Jump to content
XPEnology Community

voidru

Rookie
  • Posts

    3
  • Joined

  • Last visited

Posts posted by voidru

  1. When RAID transformation is finished I got degraded RAID5 (2 of 3 disks). 

    The new disk was marked 'crashed' so I plugged it out and made SMART check and a surface test -- no errors.

     

    Then plugged in the disk to another SATA port and started RAID5 rebuild process. 

    The rebuild finished successfully and now my DSM is healthy:

     

    image.thumb.png.beece6434917794b504206c47f43e078.png

    Thanks everyone!

    The thread can be closed. 

  2. 3 часа назад, flyride сказал:

    Is data still accessible now? 

    Yes it is.

     

    3 часа назад, flyride сказал:

    Let the RAID transformation finish, then replace the crashed drive. 

     

    Decided to follow your suggestion. Conversion will end in 24 hours. Fortunately I've made a full backup.

    • Like 1
  3. Hi, 

     

    I have the following configuration:

     

    • ESXi 6.7.0 Update 2.
    • Xpenology DSM 6.1.5-15254 (DS3615xs).
    • WD RED 2 x 3Tb added to VM using  vmkfstools -z (RDM).
    • RAID1, btfrs 

     

    I was running low on DSM, so I decided to add an additional WD Red 3 Tb (same WD model, but 4 years newer) to my VM and migrate from RAID1 to RAID5.

    I made a full backup and started the migration.

     

    The progress bar shows me that the migration will finish within 48 hours. But after several hours I got the following in the Log Center:

     

    image.thumb.png.a21d40ec09b5af0b0ba1171bb116f361.png

     

    Disk 3 (the new one) is marked as "Crashed" but RAID is coninuining changing.

     

    image.thumb.png.7bf6bcae485492309a1dcd21677635aa.png

     

    image.thumb.png.ad32972e29af1f3a0519102e063b6b62.png

     

    Here is what I have in /var/log/messages:

     

    Скрытый текст

    2020-05-17T21:44:29+03:00 DiskStation kernel: [355764.641350] md: md2: current auto_remap = 0
    2020-05-17T21:44:55+03:00 DiskStation kernel: [355790.967915] md: md2: flushing inflight I/O
    2020-05-17T21:44:55+03:00 DiskStation kernel: [355790.972967] md: reshape of RAID array md2
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855550] write error, md2, sdc3 index [2], sector 1361158280 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855557] write error, md2, sdc3 index [2], sector 1361158288 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855558] write error, md2, sdc3 index [2], sector 1361158296 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855559] write error, md2, sdc3 index [2], sector 1361158304 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855560] write error, md2, sdc3 index [2], sector 1361158312 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855561] write error, md2, sdc3 index [2], sector 1361158320 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855562] write error, md2, sdc3 index [2], sector 1361158328 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855563] write error, md2, sdc3 index [2], sector 1361158336 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855564] write error, md2, sdc3 index [2], sector 1361158344 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855565] write error, md2, sdc3 index [2], sector 1361158352 [raid5_end_write_request]
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855651] md/raid:md2: Disk failure on sdc3, disabling device.
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.855651] md/raid:md2: Operation continuing on 2 devices.
    2020-05-18T06:27:39+03:00 DiskStation kernel: [387162.858124] md: md2: reshape stop due to MD_RECOVERY_INTR set.
    2020-05-18T06:27:44+03:00 DiskStation synostoraged: hotspare_log_repair_err.c:20 [INFO] Space [/dev/md2] is actioning, skip repairing with hotspare disks
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.590787] md: md2: current auto_remap = 0
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.599042] md: md2: set sdc3 to auto_remap [1]
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.599044] md: md2: set sda3 to auto_remap [1]
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.599044] md: md2: set sdb3 to auto_remap [1]
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.599045] md: md2: flushing inflight I/O
    2020-05-18T06:27:45+03:00 DiskStation kernel: [387166.604954] md: reshape of RAID array md2


    and /proc/mdstat:

     

    Скрытый текст

     

    ash-4.3# cat /proc/mdstat
    Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1]
    md2 : active raid5 sdc3[2](F) sda3[0] sdb3[1]
          2925444544 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/2] [UU_]
          [======>..............]  reshape = 33.3% (975902464/2925444544) finish=1803.7min speed=18012K/sec

    md1 : active raid1 sdc2[2] sda2[0] sdb2[1]
          2097088 blocks [12/3] [UUU_________]

    md0 : active raid1 sdc1[12](F) sda1[0] sdb1[1]
          2490176 blocks [12/2] [UU__________]

    unused devices: <none>

     


    S.M.A.R.T on ESXi shown no errors:

     

    Скрытый текст

    [root@localhost:~] esxcli storage core device smart get -d t10.ATA_____WDC_WD30EFRX2D68EUZN0_________________________WD2DWCC4N4XTV79V
    Parameter                     Value  Threshold  Worst
    ----------------------------  -----  ---------  -----
    Health Status                 OK     N/A        N/A
    Media Wearout Indicator       N/A    N/A        N/A
    Write Error Count             0      0          N/A
    Read Error Count              0      51         N/A
    Power-on Hours                110    0          N/A
    Power Cycle Count             3      0          N/A
    Reallocated Sector Count      0      140        N/A
    Raw Read Error Rate           0      51         N/A
    Drive Temperature             46     0          N/A
    Driver Rated Max Temperature  N/A    N/A        N/A
    Write Sectors TOT Count       N/A    N/A        N/A
    Read Sectors TOT Count        N/A    N/A        N/A
    Initial Bad Block Count       N/A    N/A        N/A

     

    Nevertheless the volume is mounted and my files are accessible over SMB, but the lattency is high due to the migration.

    I believe that something is wrong with the new HDD, and I will never feel safe knowing that there were errors during RAID migration.

     

    Any suggestions?

     

    Thanks.

     

    UPD: The new disk shows 0 write\read rate in ESXi monitor:

     

    image.thumb.png.3230aa3aad88e36fdfd880573df65789.png

     

    So I guess, in the end I will get RAID5 with only 2 disks (and one is missing).

     

     

     

×
×
  • Create New...