Jump to content
XPEnology Community

Recommended Posts

Posted

Hi, 

 

I updated to DSM 6 using Jun's loader. 

 

Shortly after, I tried to add another HDD to my NAS. DSM wasn't recognizing the HDD, so I powered it off and used the SATA cables from one of the working drives to ensure it wasn't the cables. This is where everything went wrong. When DSM booted up, I saw the drive I needed, but DSM gave an error of course. After powering the unit off and swapping the cables back, it still said it needed repair, so I pressed repair in Storage Manager. Everything seemed fine. After another reboot, it said it had crashed. I let the parity check run overnight and now the RAID is running healthy as well as each individual disk, but the Volume has crashed. 

 

It's running SHR-1.

 

I believe the mdstat results below show that it can be recovered without data loss, but I'm not sure where to go from here? Another thread added the removed '/dev/sdf2' back into active sync, but I'm not sure which letters are assigned where.

 

 

/proc/mdstat/  
admin@DiskStationNAS:/usr$ cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] 
md3 : active raid1 sdc6[0] sdd6[1]
      976742784 blocks super 1.2 [2/2] [UU]
      
md2 : active raid5 sde5[0] sdd5[3] sdf5[2] sdc5[1]
      8776305792 blocks super 1.2 level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
      
md1 : active raid1 sdc2[1] sdd2[2] sde2[0] sdf2[3]
      2097088 blocks [12/4] [UUUU________]
      
md0 : active raid1 sdc1[0] sdd1[1] sde1[2] sdf1[3]
      2490176 blocks [12/4] [UUUU________]
      
unused devices: <none>
admin@DiskStationNAS:/usr$ mdadm --detail /dev/md2
mdadm: must be super-user to perform this action
admin@DiskStationNAS:/usr$ sudo mdadm --detail /dev/md2
Password: 
/dev/md2:
        Version : 1.2
  Creation Time : Sat Aug 29 05:40:53 2015
     Raid Level : raid5
     Array Size : 8776305792 (8369.74 GiB 8986.94 GB)
  Used Dev Size : 2925435264 (2789.91 GiB 2995.65 GB)
   Raid Devices : 4
  Total Devices : 4
    Persistence : Superblock is persistent

    Update Time : Sun Aug 27 09:38:44 2017
          State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

           Name : DiskStationNAS:2  (local to host DiskStationNAS)
           UUID : d97694ec:e0cb31e2:f22b36f2:86cfd4eb
         Events : 17027

    Number   Major   Minor   RaidDevice State
       0       8       69        0      active sync   /dev/sde5
       1       8       37        1      active sync   /dev/sdc5
       2       8       85        2      active sync   /dev/sdf5
       3       8       53        3      active sync   /dev/sdd5
admin@DiskStationNAS:/usr$ sudo mdadm --detail /dev/md3
/dev/md3:
        Version : 1.2
  Creation Time : Thu Jun  8 22:33:42 2017
     Raid Level : raid1
     Array Size : 976742784 (931.49 GiB 1000.18 GB)
  Used Dev Size : 976742784 (931.49 GiB 1000.18 GB)
   Raid Devices : 2
  Total Devices : 2
    Persistence : Superblock is persistent

    Update Time : Sun Aug 27 00:07:01 2017
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0

           Name : DiskStationNAS:3  (local to host DiskStationNAS)
           UUID : 4976db98:081bd234:e07be759:a005082b
         Events : 2

    Number   Major   Minor   RaidDevice State
       0       8       38        0      active sync   /dev/sdc6
       1       8       54        1      active sync   /dev/sdd6
admin@DiskStationNAS:/usr$ sudo mdadm --detail /dev/md1
/dev/md1:
        Version : 0.90
  Creation Time : Sun Aug 27 00:10:09 2017
     Raid Level : raid1
     Array Size : 2097088 (2048.28 MiB 2147.42 MB)
  Used Dev Size : 2097088 (2048.28 MiB 2147.42 MB)
   Raid Devices : 12
  Total Devices : 4
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Sun Aug 27 09:38:38 2017
          State : clean, degraded
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

           UUID : a56a9bcf:e721db67:060f5afc:c3279ded (local to host DiskStationNAS)
         Events : 0.44

    Number   Major   Minor   RaidDevice State
       0       8       66        0      active sync   /dev/sde2
       1       8       34        1      active sync   /dev/sdc2
       2       8       50        2      active sync   /dev/sdd2
       3       8       82        3      active sync   /dev/sdf2
       4       0        0        4      removed
       5       0        0        5      removed
       6       0        0        6      removed
       7       0        0        7      removed
       8       0        0        8      removed
       9       0        0        9      removed
      10       0        0       10      removed
      11       0        0       11      removed
admin@DiskStationNAS:/usr$ sudo mdadm --detail /dev/md0
/dev/md0:
        Version : 0.90
  Creation Time : Fri Dec 31 17:00:25 1999
     Raid Level : raid1
     Array Size : 2490176 (2.37 GiB 2.55 GB)
  Used Dev Size : 2490176 (2.37 GiB 2.55 GB)
   Raid Devices : 12
  Total Devices : 4
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Sun Aug 27 10:28:58 2017
          State : clean, degraded
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

           UUID : 147ea3ff:bddf1774:3017a5a8:c86610be
         Events : 0.9139

    Number   Major   Minor   RaidDevice State
       0       8       33        0      active sync   /dev/sdc1
       1       8       49        1      active sync   /dev/sdd1
       2       8       65        2      active sync   /dev/sde1
       3       8       81        3      active sync   /dev/sdf1
       4       0        0        4      removed
       5       0        0        5      removed
       6       0        0        6      removed
       7       0        0        7      removed
       8       0        0        8      removed
       9       0        0        9      removed
      10       0        0       10      removed
      11       0        0       11      removed

 

Hide  
Posted


So, my problem may be related to the disks being out of order. Checking /etc/space for an earlier .xml log, it shows they were in alphabetical order (sdc5, sdd5, sde5, sdf5). 

I tried using:

syno_poweroff_task -d 
mdadm --stop /dev/md2
mdadm -Cf /dev/md2 -e1.2 -n4 -l5 /dev/sdc5 /dev/sdd5 /dev/sde5 /dev/sdf5 -ud97694ec:e0cb31e2:f22b36f2:86cfd4eb

but it kept resulting in the error "mdadm: failed to stop array /dev/md2: Device or resource busy"

Posted (edited)
18 minutes ago, Polanskiman said:

Are you sure you didn't change the drives order when plugging them back? They need to be plugged in the exact same order as they were originally. 

 

Wouldn't I have had to make two swaps in order to get the result below?:

 Number   Major   Minor   RaidDevice State
       0       8       69        0      active sync   /dev/sde5
       1       8       37        1      active sync   /dev/sdc5
       2       8       85        2      active sync   /dev/sdf5
       3       8       53        3      active sync   /dev/sdd5

They were originally ordered c5, d5, e5, f5.

 

If that is the case, then it wasn't the physical connection of the drives that changed things because I only unplugged one drive.

Edited by redcinex
Posted

I just checked the /etc/space logs and the order currently appears alphabetically there as well. So that information I found earlier doesn't prove that the disks are out of order. :(

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...