flyride Posted April 5, 2022 Share #26 Posted April 5, 2022 Let's try a normal array start. mdadm --run /dev/md3 and cat /proc/mdstat Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #27 Posted April 5, 2022 mdadm --run /dev/md3 mdadm: failed to start array /dev/md3: Invalid argument root@NAS:~# cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1] md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9] 2097088 blocks [12/9] [UUUU_UUUUU__] md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9] 2490176 blocks [12/7] [UUU_UU__UU__] unused devices: <none> Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #28 Posted April 5, 2022 One more variant: mdadm --assemble /dev/md3 cat /proc/mdstat Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #29 Posted April 5, 2022 mdadm --assemble /dev/md3 mdadm: /dev/md3 not identified in config file. root@NAS:~# cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1] md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9] 2097088 blocks [12/9] [UUUU_UUUUU__] md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9] 2490176 blocks [12/7] [UUU_UU__UU__] unused devices: <none> Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #30 Posted April 5, 2022 That I don't understand at all. Try this: mdadm --assemble --scan cat /proc/mdstat Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #31 Posted April 5, 2022 (edited) mdadm --assemble --scan mdadm: /dev/md/3 has been started with 9 drives (out of 10). mdadm: Found some drive for an array that is already active: /dev/md/0_0 mdadm: giving up. root@NAS:~# cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1] md3 : active raid5 sda3[0] sdg3[11] sdk3[9] sdb3[10] sdl3[6] sdj3[8] sdf3[3] sdd3[2] sdc3[1] 35119473984 blocks super 1.2 level 5, 64k chunk, algorithm 2 [10/9] [UUUU_UUUUU] md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9] 2097088 blocks [12/9] [UUUU_UUUUU__] md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9] 2490176 blocks [12/7] [UUU_UU__UU__] unused devices: <none> Edited April 5, 2022 by costyy Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #32 Posted April 5, 2022 Don't worry about what Storage Manager says right now. mount and df -v Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #33 Posted April 5, 2022 mount /dev/md0 on / type ext4 (rw,relatime,journal_checksum,barrier,data=ordered) none on /dev type devtmpfs (rw,nosuid,noexec,relatime,size=16459916k,nr_inodes=4114979,mode=755) none on /dev/pts type devpts (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000) none on /proc type proc (rw,nosuid,nodev,noexec,relatime) none on /sys type sysfs (rw,nosuid,nodev,noexec,relatime) /tmp on /tmp type tmpfs (rw,relatime) /run on /run type tmpfs (rw,nosuid,nodev,relatime,mode=755) /dev/shm on /dev/shm type tmpfs (rw,nosuid,nodev,relatime) none on /sys/fs/cgroup type tmpfs (rw,relatime,size=4k,mode=755) cgmfs on /run/cgmanager/fs type tmpfs (rw,relatime,size=100k,mode=755) cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,relatime,cpuset,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuset,clone_children) cgroup on /sys/fs/cgroup/cpu type cgroup (rw,relatime,cpu,release_agent=/run/cgmanager/agents/cgm-release-agent.cpu) cgroup on /sys/fs/cgroup/cpuacct type cgroup (rw,relatime,cpuacct,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuacct) cgroup on /sys/fs/cgroup/memory type cgroup (rw,relatime,memory,release_agent=/run/cgmanager/agents/cgm-release-agent.memory) cgroup on /sys/fs/cgroup/devices type cgroup (rw,relatime,devices,release_agent=/run/cgmanager/agents/cgm-release-agent.devices) cgroup on /sys/fs/cgroup/freezer type cgroup (rw,relatime,freezer,release_agent=/run/cgmanager/agents/cgm-release-agent.freezer) cgroup on /sys/fs/cgroup/blkio type cgroup (rw,relatime,blkio,release_agent=/run/cgmanager/agents/cgm-release-agent.blkio) none on /proc/bus/usb type devtmpfs (rw,nosuid,noexec,relatime,size=16459916k,nr_inodes=4114979,mode=755) none on /sys/kernel/debug type debugfs (rw,relatime) securityfs on /sys/kernel/security type securityfs (rw,relatime) none on /config type configfs (rw,relatime) root@NAS:~# df -v Filesystem 1K-blocks Used Available Use% Mounted on /dev/md0 2385528 1624852 641892 72% / none 16459916 0 16459916 0% /dev /tmp 16475452 928 16474524 1% /tmp /run 16475452 4896 16470556 1% /run /dev/shm 16475452 4 16475448 1% /dev/shm none 4 0 4 0% /sys/fs/cgroup cgmfs 100 0 100 0% /run/cgmanager/fs Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #34 Posted April 5, 2022 ls -la / Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #35 Posted April 5, 2022 ls -la / total 100 drwxr-xr-x 30 root root 4096 Apr 5 18:30 . drwxr-xr-x 30 root root 4096 Apr 5 18:30 .. lrwxrwxrwx 1 root root 7 Aug 27 2021 bin -> usr/bin drwxr-xr-x 7 root root 0 Apr 5 18:30 config drwxr-xr-x 13 root root 19340 Apr 5 19:29 dev drwxr-xr-x 49 root root 4096 Apr 5 18:34 etc drwxr-xr-x 43 root root 4096 Aug 27 2021 etc.defaults drwxr-xr-x 2 root root 4096 May 12 2020 initrd lrwxrwxrwx 1 root root 7 Aug 27 2021 lib -> usr/lib lrwxrwxrwx 1 root root 9 Aug 27 2021 lib32 -> usr/lib32 lrwxrwxrwx 1 root root 7 Aug 27 2021 lib64 -> usr/lib drwxr-xr-x 2 root root 4096 Aug 27 2021 .log.junior drwx------ 2 root root 4096 May 12 2020 lost+found -rw-r--r-- 1 root root 0 Apr 2 00:28 Media drwxr-xr-x 2 root root 4096 May 12 2020 mnt drwxr-xr-x 2 root root 4096 Aug 27 2021 .old_patch_info drwx--x--x 4 root root 4096 Sep 2 2021 opt dr-xr-xr-x 248 root root 0 Apr 5 18:27 proc -rw------- 1 root root 1024 Aug 27 2021 .rnd drwx------ 5 root root 4096 Dec 18 13:08 root drwxr-xr-x 26 root root 1360 Apr 5 20:00 run lrwxrwxrwx 1 root root 8 Aug 27 2021 sbin -> usr/sbin -rw-r--r-- 1 root root 0 Apr 2 00:28 Server drwxr-xr-x 4 root root 4096 Aug 27 2021 .syno dr-xr-xr-x 12 root root 0 Apr 5 18:30 sys drwxr-xr-x 2 root root 4096 Aug 27 2021 .system_info drwxrwxrwt 12 root root 1440 Apr 5 20:08 tmp drwxr-xr-x 2 root root 4096 Aug 27 2021 tmpRoot drwxr-xr-x 11 root root 4096 May 12 2020 usr drwxr-xr-x 17 root root 4096 Apr 5 18:30 var drwxr-xr-x 14 root root 4096 Aug 27 2021 var.defaults drwxrwxrwx 2 root root 4096 Apr 1 14:17 volume1 drwxr-xr-x 4 root root 4096 Jan 14 20:38 volume2 drwxr-xr-x 5 root root 4096 Nov 1 20:05 volumeSATA1 drwxr-xr-x 3 root root 4096 Nov 1 19:39 volumeSATA2 drwxr-xr-x 3 root root 4096 Nov 1 19:39 volumeSATA3 drwxr-xr-x 3 root root 4096 Apr 2 00:31 volumeUSB1 drwxr-xr-x 3 root root 4096 Jan 19 09:45 volumeUSB2 Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #36 Posted April 5, 2022 mount -v /dev/md3 /volume2 Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #37 Posted April 5, 2022 (edited) mount -v /dev/md3 /volume2 mount: /dev/md3 mounted on /volume2 thank you very much how can i make it up to you Edited April 5, 2022 by costyy Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #38 Posted April 5, 2022 (edited) Ok. So this is good news. You should be able to access your shares on /volume2 normally via the network, or via File Station. BUT, your nas is not in good shape. The /dev/md3 aka /volume2 array is critical (no redundancy). Any disturbance could cause loss of some or all of your data. So please consider the following data recovery process going forward: 1. Do not reboot 2. Do not attempt to fix anything in Storage Manager 3. Copy ALL your data off the nas 4. Delete your Volume2/Storage Pool 2 5. Replace any disks with bad sectors 6. Fix the System Partition using the button in Storage Manager 7. Build a new Storage Pool (array), testing the integrity of the drives 8. Build a new btrfs volume 9. Copy your files back onto the nas Please note that none of this does anything for your SSD, which I think is /dev/sde and normally mapped to /volume1. I don't know what is wrong with it, or if you have any important data on it. DSM doesn't even believe it exists at the moment. However, it isn't really affected by what is happening with your main data array. Edited April 5, 2022 by flyride Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #39 Posted April 5, 2022 i will copy the data and delete all after that.and start from scrach Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #40 Posted April 5, 2022 (edited) the volume crushed again and the speed of trasnfer is 1000kb and all off the data is rubish.cam i do the steps again to unmount and remount the volume Edited April 5, 2022 by costyy Quote Link to comment Share on other sites More sharing options...
flyride Posted April 5, 2022 Share #41 Posted April 5, 2022 If there is a drive retrying a bad sector for a long time, the transfer rate will be slow while those sectors are being encountered. That is what you want (the drive trying to recover the data). Usually DSM will retry and then fail a particular file. But it might decide to crash the array if the drive error is significant enough. Are all the member drives still visible or has one or more disks gone offline? If drives are offline, that means you have some sort of continued hardware problem. Either a cable, controller or drive is failing. Nobody can diagnose the issue without you posting more information about the crash and circumstances. This needs to be fixed before trying anything again. If a drive is unusable at this point, your data is incomplete. You could send the array to a forensic recovery specialist at great cost, and they will retrieve some data from the array, but not all of it. "All the data is rubbish." Based on the signatures of the drives, this should not be the case, unless you modified one or more of the drives while DSM was not in control of them. I can't advise you on that. There will be some minor corruption based on the stale drive I talked about earlier but it should only be a few files if any. Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #42 Posted April 5, 2022 on all 10 drives i have 3 bad sectors 1 on 3 drives,i transfered some data from the volume and all is not usable .the status of volume 2 changed from degraded to crushed.now the drive 7 is crushed Quote Link to comment Share on other sites More sharing options...
costyy Posted April 5, 2022 Author Share #43 Posted April 5, 2022 how to unmount the array and mount it again Quote Link to comment Share on other sites More sharing options...
flyride Posted April 6, 2022 Share #44 Posted April 6, 2022 A simple bad sector won't cause the problems you seem to be having. I think you have a power problem or a cable problem. Please try and resolve this first before trying to restart the array again. The information you need to investigate and recover the array just like we have done is in this thread now. Post if you need help. Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.