Chuck12 Posted February 24, 2021 Share #1 Posted February 24, 2021 I read the other post by pavey but my situation is different. One of my pool crashed and the actions button is grayed out. The volume is a SHR-2. I followed Synology's suggestion to recover the data from https://www.synology.com/en-us/knowledgebase/DSM/tutorial/Storage/How_can_I_recover_data_from_my_DiskStation_using_a_PC but got stuck at step 9. I get the following when I run mdadm -Asf && vgchange -ay mdadm: No arrays found in config file or automatically I launched gparted and can see the volume info /dev/md127 14.44Tib Partition: /devmd127 FileSystem: lvm2 pv MountPoint: vg1003 Size: 14.44TiB /dev/sda5/7.22TiB /dev/sdb5/7.22TiB /dev/sdc5/7.22TiB Help! Any suggestions? Thank you! Quote Link to comment Share on other sites More sharing options...
flyride Posted February 24, 2021 Share #2 Posted February 24, 2021 This guide doesn't help you recover any data; it only gets you access to a functioning array if you can't boot DSM. If you have a problem with your array, it very well will make things worse. Put it all back and start posting diagnostic information. @pdavey's approach is a good start. Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted February 24, 2021 Author Share #3 Posted February 24, 2021 I put the drives back and the repair option re-appeared. One of the drive was bad so I'll need to get a replacement drive in its place to do the repair. Thanks Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted February 28, 2021 Author Share #4 Posted February 28, 2021 The repair process completed, drives are shown as healthy, storage pool shows healthy but volume shows crashed. This is for Storage Pool 4 / Volume 4. I'm running xpenology in VMWare Workstation 16. root@NAS:~# cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1] md5 : active raid5 sde5[0] sdd5[3] sdf5[1] 9637230720 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md3 : active raid1 sdp5[0] 1882600064 blocks super 1.2 [1/1] [U] md2 : active raid5 sdj5[3] sdl5[5] sdk5[4] 15509261312 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md4 : active raid5 sdi5[0] sdh5[2] sdg5[1] 7749802112 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md1 : active raid1 sdp2[9] sdl2[8] sdk2[7] sdj2[6] sdi2[5] sdh2[4] sdg2[3] sdf2[2] sde2[1] sdd2[0] 2097088 blocks [24/10] [UUUUUUUUUU______________] md0 : active raid1 sdd1[0] sde1[1] sdf1[2] sdg1[4] sdh1[6] sdi1[5] sdj1[3] sdk1[8] sdl1[7] sdp1[9] 2490176 blocks [12/10] [UUUUUUUUUU__] unused devices: <none> mdadm --detail /dev/md2 /dev/md2: Version : 1.2 Creation Time : Wed Feb 5 00:22:58 2020 Raid Level : raid5 Array Size : 15509261312 (14790.78 GiB 15881.48 GB) Used Dev Size : 7754630656 (7395.39 GiB 7940.74 GB) Raid Devices : 3 Total Devices : 3 Persistence : Superblock is persistent Update Time : Sat Feb 27 16:43:49 2021 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 64K Name : Plex:2 UUID : f9ac4c36:6d621467:46a38f36:e75e077a Events : 52639 Number Major Minor RaidDevice State 3 8 149 0 active sync /dev/sdj5 4 8 165 1 active sync /dev/sdk5 5 8 181 2 active sync /dev/sdl5 root@NAS:~# fdisk -l /dev/sdj5 Disk /dev/sdj5: 7.2 TiB, 7940742873088 bytes, 15509263424 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes root@NAS:~# fdisk -l /dev/sdk5 Disk /dev/sdk5: 7.2 TiB, 7940742873088 bytes, 15509263424 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes root@NAS:~# fdisk -l /dev/sdl5 Disk /dev/sdl5: 7.2 TiB, 7940742873088 bytes, 15509263424 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes root@NAS:~# vgchange -ay 1 logical volume(s) in volume group "vg1000" now active 1 logical volume(s) in volume group "vg1001" now active 1 logical volume(s) in volume group "vg1002" now active 1 logical volume(s) in volume group "vg1003" now active root@NAS:~# cat /etc/fstab none /proc proc defaults 0 0 /dev/root / ext4 defaults 1 1 /dev/vg1003/lv /volume4 btrfs auto_reclaim_space,synoacl,relatime 0 0 /dev/vg1002/lv /volume1 btrfs auto_reclaim_space,synoacl,ssd,relatime 0 0 /dev/vg1000/lv /volume3 btrfs auto_reclaim_space,synoacl,relatime 0 0 /dev/vg1001/lv /volume2 btrfs auto_reclaim_space,synoacl,relatime 0 0 root@NAS:~# lvm pvscan PV /dev/md5 VG vg1000 lvm2 [8.98 TiB / 0 free] PV /dev/md4 VG vg1001 lvm2 [7.22 TiB / 0 free] PV /dev/md3 VG vg1002 lvm2 [1.75 TiB / 0 free] PV /dev/md2 VG vg1003 lvm2 [14.44 TiB / 0 free] Total: 4 [32.39 TiB] / in use: 4 [32.39 TiB] / in no VG: 0 [0 ] Please let me know what other details you need. Thanks in advance! Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted February 28, 2021 Author Share #5 Posted February 28, 2021 After lots of googling, I was able to get access to my data from that crashed volume. I removed the vhd from the VM, added them to an ubuntu VM and followed the Synology link to recover the data. This was done after the raid was repaired that it showed up in the ubuntu files manager app. Previously when the failed drive was not repaired, the volume was not available. It was visible as an option but an error appeared when I tried to access it. I ran the lvscan to display what to mount and mine showed /dev/vg1003/lv. Since it was a btrfs file system, I then installed the btrfs-progs package and ran the following mount command: mount -t btrfs -o ro,recovery /dev/vg1003/lv /mnt/disk. For other linux noobs like myself, I had already created a /mnt/disk earlier so remember to do that if you had not. I waited for about 10 seconds and the mount command completed. Went back to Files manager and the volume disappeared. I thought CRAP! but then I just remember that I had mounted it in /mnt/disk. Sure enough, opened /mnt/disk and the folders are there! 1 Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted March 10, 2021 Author Share #6 Posted March 10, 2021 Well that didn't last long. A windows update forced a reboot while the nas was doing a consistency check on a separate volume. Unfortunately, this time, I can't mount the array in ubuntu to access the data. Below are the diagnostic data of the drives in the nas. Note that two of the missing drives are shown as Initialized and healthy in the HDD/SSD group but they are not assigned to the specific pool. HELP! root@NAS:~# cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1] md4 : active raid5 sdi5[0] sdo5[3] 7749802112 blocks super 1.2 level 5, 64k chunk, algorithm 2 [4/2] [U__U] md3 : active raid1 sdp5[0] 1882600064 blocks super 1.2 [1/1] [U] md5 : active raid5 sde5[0] sdd5[3] sdf5[1] 9637230720 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md6 : active raid5 sdj6[0] sdl6[2] sdk6[1] 13631455616 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md2 : active raid5 sdj5[0] sdl5[3] sdk5[1] 1877771392 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU] md7 : active raid1 sdi6[0] sdo6[1] 838848960 blocks super 1.2 [2/2] [UU] md1 : active raid1 sdp2[11] sdo2[10] sdn2[9] sdl2[8] sdk2[7] sdj2[6] sdi2[5] sdh2[4] sdg2[3] sdf2[2] sde2[1] sdd2[0] 2097088 blocks [24/12] [UUUUUUUUUUUU____________] md0 : active raid1 sdd1[0] sde1[1] sdf1[9] sdg1[8] sdh1[3] sdi1[11] sdj1[10] sdk1[5] sdl1[4] sdn1[7] sdo1[6] sdp1[2] 2490176 blocks [12/12] [UUUUUUUUUUUU] unused devices: <none> Odd I only have a total of 4 Volumes? root@NAS:~# mdadm --detail /dev/md4 /dev/md4: Version : 1.2 Creation Time : Sat Feb 1 18:20:02 2020 Raid Level : raid5 Array Size : 7749802112 (7390.79 GiB 7935.80 GB) Used Dev Size : 3874901056 (3695.39 GiB 3967.90 GB) Raid Devices : 4 Total Devices : 2 Persistence : Superblock is persistent Update Time : Tue Mar 9 08:12:16 2021 State : clean, FAILED Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 64K Delta Devices : 1, (3->4) Name : NAS:4 (local to host NAS) UUID : f95f4a28:1cd4a412:0df847fe:4645d62b Events : 8456 Number Major Minor RaidDevice State 0 8 133 0 active sync /dev/sdi5 - 0 0 1 removed - 0 0 2 removed 3 8 229 3 active sync /dev/sdo5 root@NAS:~# fdisk -l /dev/sdi5 Disk /dev/sdi5: 3.6 TiB, 3967899729920 bytes, 7749804160 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes root@NAS:~# fdisk -l /dev/sdo5 Disk /dev/sdo5: 3.6 TiB, 3967899729920 bytes, 7749804160 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes root@NAS:~# pvs /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. PV VG Fmt Attr PSize PFree /dev/md2 vg1003 lvm2 a-- 1.75t 0 /dev/md3 vg1002 lvm2 a-- 1.75t 0 /dev/md5 vg1000 lvm2 a-- 8.98t 0 /dev/md6 vg1003 lvm2 a-- 12.70t 0 /dev/md7 vg1001 lvm2 a-- 799.98g 0 unknown device vg1001 lvm2 a-m 7.22t 0 root@NAS:~# vgs /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. VG #PV #LV #SN Attr VSize VFree vg1000 1 1 0 wz--n- 8.98t 0 vg1001 2 1 0 wz-pn- 8.00t 0 vg1002 1 1 0 wz--n- 1.75t 0 vg1003 2 1 0 wz--n- 14.44t 0 root@NAS:~# lvs /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert lv vg1000 -wi-ao---- 8.98t lv vg1001 -wi-----p- 8.00t lv vg1002 -wi-ao---- 1.75t lv vg1003 -wi-ao---- 14.44t root@NAS:~# pvdisplay /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. --- Physical volume --- PV Name unknown device VG Name vg1001 PV Size 7.22 TiB / not usable 2.12 MiB Allocatable yes (but full) PE Size 4.00 MiB Total PE 1892041 Free PE 0 Allocated PE 1892041 PV UUID eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F --- Physical volume --- PV Name /dev/md7 VG Name vg1001 PV Size 799.99 GiB / not usable 4.44 MiB Allocatable yes (but full) PE Size 4.00 MiB Total PE 204796 Free PE 0 Allocated PE 204796 PV UUID E9ibKv-ZC3q-yWDp-l0C6-bxlf-5Kvq-SFNh7g --- Physical volume --- PV Name /dev/md5 VG Name vg1000 PV Size 8.98 TiB / not usable 1.56 MiB Allocatable yes (but full) PE Size 4.00 MiB Total PE 2352839 Free PE 0 Allocated PE 2352839 PV UUID PfIiwF-d5d5-t46T-7a0W-TMjL-4nnh-SvRJQF --- Physical volume --- PV Name /dev/md3 VG Name vg1002 PV Size 1.75 TiB / not usable 640.00 KiB Allocatable yes (but full) PE Size 4.00 MiB Total PE 459619 Free PE 0 Allocated PE 459619 PV UUID JLchK0-rzk8-lYZd-s2dR-tiAE-xFhq-10cE8q --- Physical volume --- PV Name /dev/md2 VG Name vg1003 PV Size 1.75 TiB / not usable 0 Allocatable yes (but full) PE Size 4.00 MiB Total PE 458440 Free PE 0 Allocated PE 458440 PV UUID U44r7Z-5MC3-nHMi-dwFA-gYtD-V0mq-tve2RK --- Physical volume --- PV Name /dev/md6 VG Name vg1003 PV Size 12.70 TiB / not usable 3.81 MiB Allocatable yes (but full) PE Size 4.00 MiB Total PE 3327991 Free PE 0 Allocated PE 3327991 PV UUID uxtEKL-6X6d-1QYL-2mny-kvSe-X3fV-qPpesk root@NAS:~# vgdisplay /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. --- Volume group --- VG Name vg1001 System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 4 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 0 Max PV 0 Cur PV 2 Act PV 1 VG Size 8.00 TiB PE Size 4.00 MiB Total PE 2096837 Alloc PE / Size 2096837 / 8.00 TiB Free PE / Size 0 / 0 VG UUID 3ZKEpX-f4o2-8Zc8-AXax-NAC2-xEGs-BgQRzp --- Volume group --- VG Name vg1000 System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 4 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 1 Max PV 0 Cur PV 1 Act PV 1 VG Size 8.98 TiB PE Size 4.00 MiB Total PE 2352839 Alloc PE / Size 2352839 / 8.98 TiB Free PE / Size 0 / 0 VG UUID aHQHQk-zx96-OWTT-tjK0-hIy4-up1V-4DrBq3 --- Volume group --- VG Name vg1002 System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 2 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 1 Max PV 0 Cur PV 1 Act PV 1 VG Size 1.75 TiB PE Size 4.00 MiB Total PE 459619 Alloc PE / Size 459619 / 1.75 TiB Free PE / Size 0 / 0 VG UUID 7GhuzZ-Ba7k-juDK-X1Ez-C6Xk-Dceo-K0SiEa --- Volume group --- VG Name vg1003 System ID Format lvm2 Metadata Areas 2 Metadata Sequence No 5 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 1 Max PV 0 Cur PV 2 Act PV 2 VG Size 14.44 TiB PE Size 4.00 MiB Total PE 3786431 Alloc PE / Size 3786431 / 14.44 TiB Free PE / Size 0 / 0 VG UUID ga4uN9-NL9S-o9Cr-uK1L-nENj-9Mcw-71toaL root@NAS:~# lvdisplay /dev/md4: read failed after 0 of 4096 at 0: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797297152: Input/output error /dev/md4: read failed after 0 of 4096 at 7935797354496: Input/output error /dev/md4: read failed after 0 of 4096 at 4096: Input/output error Couldn't find device with uuid eJGcET-fSVZ-JCMi-kaHL-dipI-G4s6-r1R32F. --- Logical volume --- LV Path /dev/vg1001/lv LV Name lv VG Name vg1001 LV UUID 3ZaAe2-Bs3y-0G1k-hr9O-OSIj-F0XN-12COkF LV Write Access read/write LV Creation host, time , LV Status NOT available LV Size 8.00 TiB Current LE 2096837 Segments 2 Allocation inherit Read ahead sectors auto --- Logical volume --- LV Path /dev/vg1000/lv LV Name lv VG Name vg1000 LV UUID R5pH8E-X9AL-2Afw-OPsq-nOpy-ffHO-CCvWfp LV Write Access read/write LV Creation host, time , LV Status available # open 1 LV Size 8.98 TiB Current LE 2352839 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 4096 Block device 253:0 --- Logical volume --- LV Path /dev/vg1002/lv LV Name lv VG Name vg1002 LV UUID KnvT1S-9R4B-wFui-kYnQ-i9ns-BGQP-yotSts LV Write Access read/write LV Creation host, time , LV Status available # open 1 LV Size 1.75 TiB Current LE 459619 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 4096 Block device 253:2 --- Logical volume --- LV Path /dev/vg1003/lv LV Name lv VG Name vg1003 LV UUID c4intm-bTUc-rC4n-jI2q-f4nv-st63-Uz5ze3 LV Write Access read/write LV Creation host, time , LV Status available # open 1 LV Size 14.44 TiB Current LE 3786431 Segments 3 Allocation inherit Read ahead sectors auto - currently set to 4096 Block device 253:1 root@NAS:~# cat /etc/fstab none /proc proc defaults 0 0 /dev/root / ext4 defaults 1 1 /dev/vg1001/lv /volume2 btrfs 0 0 /dev/vg1002/lv /volume1 btrfs auto_reclaim_space,synoacl,ssd,relatime 0 0 /dev/vg1003/lv /volume4 btrfs auto_reclaim_space,synoacl,ssd,relatime 0 0 /dev/vg1000/lv /volume3 btrfs auto_reclaim_space,synoacl,relatime 0 0 Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 12, 2021 Share #7 Posted March 12, 2021 On 3/10/2021 at 1:34 AM, Chuck12 said: Well that didn't last long. A windows update forced a reboot while the nas was doing a consistency check on a separate volume. Unfortunately, this time, I can't mount the array in ubuntu to access the data you run it as a virtual box vm on windows? if so you should check that host i/o caching is switched off, if its on you might loose important data when virtualbox is abruptly forced to end and that might be the cause of your problems Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted March 12, 2021 Author Share #8 Posted March 12, 2021 it's definitely off Quote Link to comment Share on other sites More sharing options...
Mincemeat Posted March 31, 2021 Share #9 Posted March 31, 2021 Why u no gpedit the shity windows so it never forces updates? Did u manage to recover? Quote Link to comment Share on other sites More sharing options...
Chuck12 Posted March 31, 2021 Author Share #10 Posted March 31, 2021 I have disabled it since the host redo. there were a few things that caused the problem. one of the drives in the pool was bad which caused vmware to halt during a scrub. another was a kernel power failure from a bad windows update. a clean install of windows so far have been stable. I have the vhds set aside. still have not recovered yet. need to read up on Linux some more. I thought I did everything from my first attempt but it's not working second time. Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.