• Content Count

  • Joined

  • Last visited

Community Reputation

0 Neutral

About sarieri

  • Rank
    Junior Member

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. Hi, Sorry to bother again, but it looks there is still some problem. So I played with SataPortmap in grub.cfg in order to have all six drives that connect to the AHCI controller recognized by DSM. DSM did recognize all 6 drives, but it performs a repair on the storage pool(SHR). The Storage pool is repaired now and shows healthy, but the volume is creashed and data is read only. Do you know if there's anything I can do now or I just have to backup things and rebuild everything.
  2. Now this is interesting. I tried 3617 with jun's loader 1.03 and it works. I have no idea why the 918+ works differently.
  3. Not working...Well, I did some research on google and maybe I will give it a try later. For now, KVM and UNRAID passthru the AHCI controller fine but not ESXI. Might be something wrong with reset mode(d3d0) of the controller.
  4. Should that be added here ? set extra_args_918='intel_iommu=on iommu=soft'
  5. This is weird, the controller itself must be okay since I passed through them to DSM in UNRAID before. So this might be something related to ESXI.
  6. So this has nothing to do with the configuration of set extra_args_918 in grub.cfg right?
  7. Hi, Here is the full version dmesg...I thought that vim could pull out the full bad. [Sun Jun 28 01:24:51 2020] pci 0000:13:00.0: Signaling PME through PCIe PME interrupt ..... [Sun Jun 28 01:24:52 2020] ahci 0000:13:00.0: AHCI 0001.0300 32 slots 6 ports 6 Gbps 0x3f impl SATA mode [Sun Jun 28 01:24:52 2020] ahci 0000:13:00.0: flags: 64bit ncq led clo pio slum part ems apst [Sun Jun 28 01:24:52 2020] scsi host4: ahci [Sun Jun 28 01:24:52 2020] scsi host5: ahci [Sun Jun 28 01:24:52 2020] scsi host6: ahci [Sun Jun 28 01:24:52 2020] scsi host7: ahci [Sun Jun 28 01:24:52 2
  8. Does "failed to assign" meant that there's no space for the ahci controller anymore so that I need to edit the "set extra_args_918" line in grub.cfg? I looked into disk mapping issues a while ago but never really understand how this actually work... It seems to me that there are only two controllers for DSM in my case. One is the virtual controller connected to the disk with loader and other two RDM passthru NVME drives, the other controller is the passed through AHCI controller.
  9. Sarieri@Sarieri:/$ lspci -k 0000:00:00.0 Class 0600: Device 8086:7190 (rev 01) Subsystem: Device 15ad:1976 Kernel driver in use: agpgart-intel 0000:00:01.0 Class 0604: Device 8086:7191 (rev 01) 0000:00:07.0 Class 0601: Device 8086:7110 (rev 08) Subsystem: Device 15ad:1976 0000:00:07.1 Class 0101: Device 8086:7111 (rev 01) Subsystem: Device 15ad:1976 0000:00:07.3 Class 0680: Device 8086:7113 (rev 08) Subsystem: Device 15ad:1976 0000:00:07.7 Class 0880: Device 15ad:0740 (rev 10) Subsystem: Device 15ad:0740 0000:00:0f.0 Class 0300: Device 15ad:0405
  10. Is it possible that Xpenology has the driver for the controller but I will need to edit the grub.cfg and change SataPortMap and DiskIdxMap so that it can recognize it properly.
  11. Hi, Do you know if dsm918+ have driver for the onboard SATA controller? I have a Supermicro X10SRL-F with two onboard Wellsburg AHCI Controllers. They were fine passsing through to Xpenology in KVM but not in ESXI. I'm kind of lost whether it's a driver issue of xpenology or something wrong with ESXI. I tried to passthrough the Wellsburg AHCI controller to windows in ESXI and the disks were recognized without problem...make the situation even more wired. I guess it might be something wrong with the reset method I set in in ESXI. The Wellsburg AHCI Controller was or
  12. Thank you so much for clearification.
  13. Yes, I understand that but before applying the NVME patch, synology recognize nvme drives as normal drives only available to use as storage drives. So, I'm just wondering if there's a way to remove the patch.
  14. Does anyone know how to remove this nvme patch so that one can use the nvme drive as storage again?