petersnows

One or more raid groups/ssd caches are crashed

Recommended Posts

 

One or more raid groups/ssd caches are crashed.
I have a volume 3 that shows up as crashed.
It had been working fine for more than a week

I can accees it though (all the files are fine).
I would like to know how do I change the HD/volume/raidgroup  status.
I actually don't know why it shows as crashed.


hp microserver gen8
ESXI 6.5
3TB Hardrive, RDM, (Raw Device Mapping)
6.1.15047
Jun 1.02b
 

** Mdstat


root@fs02:/etc/space# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1]
md2 : active raid1 sdb5[0]
      47695808 blocks super 1.2 [1/1] [U]

md3 : active raid1 sdc3[0](E)
      2925444544 blocks super 1.2 [1/1] [E]

md1 : active raid1 sdc2[1] sdb2[0]
      2097088 blocks [12/2] [UU__________]

md0 : active raid1 sdc1[1] sdb1[0]
      2490176 blocks [12/2] [UU__________]

unused devices: <none>
root@fs02:/etc/space#

** mdadm


root@fs02:/etc/space# mdadm --detail /dev/md3
/dev/md3:
        Version : 1.2
  Creation Time : Sun Apr 16 19:58:33 2017
     Raid Level : raid1
     Array Size : 2925444544 (2789.92 GiB 2995.66 GB)
  Used Dev Size : 2925444544 (2789.92 GiB 2995.66 GB)
   Raid Devices : 1
  Total Devices : 1
    Persistence : Superblock is persistent

    Update Time : Tue Jun 27 23:15:13 2017
          State : clean
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0

           Name : fs03:3
           UUID : bbe73c5c:694d1437:20f54790:be2b9bfb
         Events : 5

    Number   Major   Minor   RaidDevice State
       0       8       35        0      active sync   /dev/sdc3
root@fs02:/etc/space#

** mount


root@fs02:/etc/space# mount
/dev/md0 on / type ext4 (rw,relatime,journal_checksum,barrier,data=ordered)
none on /dev type devtmpfs (rw,nosuid,noexec,relatime,size=1022480k,nr_inodes=255620,mode=755)
none on /dev/pts type devpts (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000)
none on /proc type proc (rw,nosuid,nodev,noexec,relatime)
none on /sys type sysfs (rw,nosuid,nodev,noexec,relatime)
/tmp on /tmp type tmpfs (rw,relatime)
/run on /run type tmpfs (rw,nosuid,nodev,relatime,mode=755)
/dev/shm on /dev/shm type tmpfs (rw,nosuid,nodev,relatime)
none on /sys/fs/cgroup type tmpfs (rw,relatime,size=4k,mode=755)
cgmfs on /run/cgmanager/fs type tmpfs (rw,relatime,size=100k,mode=755)
cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,relatime,cpuset,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuset,clone_children)
cgroup on /sys/fs/cgroup/cpu type cgroup (rw,relatime,cpu,release_agent=/run/cgmanager/agents/cgm-release-agent.cpu)
cgroup on /sys/fs/cgroup/cpuacct type cgroup (rw,relatime,cpuacct,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuacct)
cgroup on /sys/fs/cgroup/memory type cgroup (rw,relatime,memory,release_agent=/run/cgmanager/agents/cgm-release-agent.memory)
cgroup on /sys/fs/cgroup/devices type cgroup (rw,relatime,devices,release_agent=/run/cgmanager/agents/cgm-release-agent.devices)
cgroup on /sys/fs/cgroup/freezer type cgroup (rw,relatime,freezer,release_agent=/run/cgmanager/agents/cgm-release-agent.freezer)
cgroup on /sys/fs/cgroup/blkio type cgroup (rw,relatime,blkio,release_agent=/run/cgmanager/agents/cgm-release-agent.blkio)
none on /proc/bus/usb type devtmpfs (rw,nosuid,noexec,relatime,size=1022480k,nr_inodes=255620,mode=755)
none on /sys/kernel/debug type debugfs (rw,relatime)
securityfs on /sys/kernel/security type securityfs (rw,relatime)
/dev/mapper/vol1-origin on /volume1 type ext4 (rw,relatime,journal_checksum,synoacl,data=writeback,jqfmt=vfsv0,usrjquota=aquota.user,grpjquota=aquota.group)
/dev/md3 on /volume3 type btrfs (rw,relatime,synoacl,nospace_cache,flushoncommit_threshold=1000,metadata_ratio=50)
none on /config type configfs (rw,relatime)
/dev/mapper/vol1-origin on /opt type ext4 (rw,relatime,journal_checksum,synoacl,data=writeback,jqfmt=vfsv0,usrjquota=aquota.user,grpjquota=aquota.group)
none on /proc/fs/nfsd type nfsd (rw,relatime)
/dev/mapper/vol1-origin on /volume1/@docker/aufs type ext4 (rw,relatime,journal_checksum,synoacl,data=writeback,jqfmt=vfsv0,usrjquota=aquota.user,grpjquota=aquota.group)
root@fs02:/etc/space#

** console logs

[91206.299324] ata3.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0
[91206.301555] ata3.00: irq_stat 0x40000000
[91206.303097] ata3.00: failed command: WRITE DMA
[91206.304644] ata3.00: cmd ca/00:08:80:06:4c/00:00:00:00:00/e0 tag 28 dma 4096 out
[91206.304644]          res 41/02:00:00:00:00/00:00:00:00:00/00 Emask 0x1 (device error)
[91206.309667] ata3.00: status: { DRDY ERR }
[91206.699074] ata3.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0
[91206.701293] ata3.00: irq_stat 0x40000000
[91206.702640] ata3.00: failed command: WRITE DMA
[91206.704173] ata3.00: cmd ca/00:08:80:06:4c/00:00:00:00:00/e0 tag 29 dma 4096 out
[91206.704173]          res 41/02:00:00:00:00/00:00:00:00:00/00 Emask 0x1 (device error)
[91206.709172] ata3.00: status: { DRDY ERR }

 



 

Share this post


Link to post
Share on other sites

So far I was able to resolve this by doing the following:

   - syno_poweroff_task -d
   - ssh again
   - mdadm --detail /dev/md3
     GET the UUID, mine is : bbe73c5c:694d1437:20f54790:be2b9bfb
   
        root@fs02:~# mdadm --detail /dev/md3
        /dev/md3:
                Version : 1.2
          Creation Time : Sun Apr 16 19:58:33 2017
             Raid Level : raid1
             Array Size : 2925444544 (2789.92 GiB 2995.66 GB)
          Used Dev Size : 2925444544 (2789.92 GiB 2995.66 GB)
           Raid Devices : 1
          Total Devices : 1
            Persistence : Superblock is persistent
        
            Update Time : Wed Jun 28 08:09:20 2017
                  State : clean
         Active Devices : 1
        Working Devices : 1
         Failed Devices : 0
          Spare Devices : 0
        
                   Name : fs03:3
                   UUID : bbe73c5c:694d1437:20f54790:be2b9bfb
                 Events : 5
        
            Number   Major   Minor   RaidDevice State
               0       8       35        0      active sync   /dev/sdc3
        root@fs02:~#
   
   - mdadm --stop /dev/md3         
   - btrfs check --repair /dev/md3
   - mdadm -Cf /dev/md3 -e1.2 -n1 -l1 /dev/sdc3 -ubbe73c5c:694d1437:20f54790:be2b9bfb
   - reboot

 Commands to check the status:

  cat /proc/mdstat
  mdadm --detail /dev/md3
  mdadm --examine /dev/sdc3


So far this issue happened twice.
It has been ok in the past 2 days. Let's see how it goes.

Share this post


Link to post
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.