IG-88 Posted February 25, 2018 Share #26 Posted February 25, 2018 damn it and i thought i was off the hook with the drivers Quote Link to comment Share on other sites More sharing options...
test4321 Posted March 8, 2018 Author Share #27 Posted March 8, 2018 Got my cards, but cable got stuck in shipping I'll update when I get the whole thing setup. I cant seem to find the new loader - can anybody please point me to where I can find the link to @quicknick loader? Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 8, 2018 Share #28 Posted March 8, 2018 he has not much time and wants to mod it further to incorporate all 3 versions of jun's loader so it will take a little longer but don't worry, if you really need that 10G driver i will add it shortly to my extra.lzma and you can go on with jun's loader for now 1 Quote Link to comment Share on other sites More sharing options...
test4321 Posted March 8, 2018 Author Share #29 Posted March 8, 2018 1 minute ago, IG-88 said: he has not much time and wants to mod it further to incorporate all 3 versions of jun's loader so it will take a little longer but don't worry, if you really need that 10G driver i will add it shortly to my extra.lzma and you can go on with jun's loader for now Sweet! Hopefully my ebay seller comes through and hurries the shipping up. Thanks! Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 11, 2018 Share #30 Posted March 11, 2018 i have updated the driver pack to v4.6 it now contains your brocade 10G driver (bna.ko) 1 1 Quote Link to comment Share on other sites More sharing options...
flyride Posted March 17, 2018 Share #31 Posted March 17, 2018 On 1/13/2018 at 2:58 AM, test4321 said: Was reviewing this thread and saw discussion that Mellanox Connect-X 2 might not be supported by the Synology driverset. I can confirm that the standard Mellanox driver supports Connect-X 2 single and dual port 10GBe on baremetal Xpenology 6.2 with no problem, tested on my own system. However, I switched to a ConnectX-3 because PCI 3.0 and SR-IOV support for ESXi. On 1/13/2018 at 2:58 AM, test4321 said: Hey guys, Now for networking I want to go 10GBe SFP+ I am looking at: 2X - Mellanox ConnectX-2 https://www.ebay.com/itm/391459428428 Quote Link to comment Share on other sites More sharing options...
flyride Posted March 17, 2018 Share #32 Posted March 17, 2018 On 1/13/2018 at 2:58 AM, test4321 said: Was reviewing this thread and saw discussion that Mellanox Connect-X 2 might not be supported by the Synology driverset. I can confirm that the standard Mellanox driver supports Connect-X 2 single and dual port 10GBe on baremetal Xpenology 6.2 with no problem, tested on my own system. However, I switched to a ConnectX-3 because PCI 3.0 and SR-IOV support for ESXi. On 1/13/2018 at 2:58 AM, test4321 said: Hey guys, Now for networking I want to go 10GBe SFP+ I am looking at: 2X - Mellanox ConnectX-2 https://www.ebay.com/itm/391459428428 Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 17, 2018 Share #33 Posted March 17, 2018 nice you found this but a little text would be nice too 1 Quote Link to comment Share on other sites More sharing options...
Xepnewbie2018 Posted March 17, 2018 Share #34 Posted March 17, 2018 FYI....for testing purposes only.... I bought (2) MNPA19-XTR 10GB MELLANOX CONNECTX-2 PCIe X8 10Gbe SFP+ NETWORK CARD W/CABLE from eBay.....pretty cheap ($48.00)including (2) SFP+ cables and it worked out of the box. Im using an older motherboard (MSI G41TM-E43) with only 2GB of RAM on a LGA775 Core 2Duo and Im getting 398.7 Write and 452.4 read when transferring files from my Mac Pro 3,1....the Mellanox doesn't work with my Mac Pro so I had to get a SolarFlare SFN5122F Dual Port 10Gbe PCIe Adapter SF329-9021-R6 ($44.00). Setup was straight forward and speeds are very close to my OWC Thunderbay IV in RAID 0 My current setup only have 4 SATAII ports with transfer rate up to 3Gb/s, ( 4 x 3TB Drives) so maybe once I move to a faster motherboard, speeds will increase. Can anyone comment or share what could contribute for speed bottleneck as the numbers of drives start to increase? 8 drives, 12 drives etc...my goal is to build a 12 drive system (36TB) 1 Quote Link to comment Share on other sites More sharing options...
test4321 Posted March 17, 2018 Author Share #35 Posted March 17, 2018 (edited) 17 minutes ago, Xepnewbie2018 said: FYI....for testing purposes only.... I bought (2) MNPA19-XTR 10GB MELLANOX CONNECTX-2 PCIe X8 10Gbe SFP+ NETWORK CARD W/CABLE from eBay.....pretty cheap ($48.00)including (2) SFP+ cables and it worked out of the box. Im using an older motherboard (MSI G41TM-E43) with only 2GB of RAM on a LGA775 Core 2Duo and Im getting 398.7 Write and 452.4 read when transferring files from my Mac Pro 3,1....the Mellanox doesn't work with my Mac Pro so I had to get a SolarFlare SFN5122F Dual Port 10Gbe PCIe Adapter SF329-9021-R6 ($44.00). Setup was straight forward and speeds are very close to my OWC Thunderbay IV in RAID 0 My current setup only have 4 SATAII ports with transfer rate up to 3Gb/s, ( 4 x 3TB Drives) so maybe once I move to a faster motherboard, speeds will increase. Can anyone comment or share what could contribute for speed bottleneck as the numbers of drives start to increase? 8 drives, 12 drives etc...my goal is to build a 12 drive system (36TB) Motherboard change is definitely the first thing to do. They are the dirt cheap part of the build - the more expensive stuff is RAM and CPU. You can probably try a build like mine - LGA1151 is cheap on eBay because of the whole Intel fuckup where they changed the socket. As far as hard drives - I think you are probably better off just buying 2 SSD drives and putting them as WRITE CACHE instead of getting 12 drives for speed. I dont know if anybody attempted to do this on XPE though. Also Synology WRITE CACHE is suspect - i have seen conflicting videos where it does improve the speed and where it doesnt at all. EDIT: I also noticed Synology started to use RAM as fast storage for Synology device databases. This is applied to Universal Search and something else (I dont remember right now). So maybe in future they will just use RAM drive for every application? Edited March 17, 2018 by test4321 Quote Link to comment Share on other sites More sharing options...
test4321 Posted March 17, 2018 Author Share #36 Posted March 17, 2018 my cable is still in transit - eBay seller DID NOT ship the cable (forgot)...DAMN IT! So probably next week I'll be able to test it. We'll see. Thanks @IG-88 for your work!!! Quote Link to comment Share on other sites More sharing options...
Xepnewbie2018 Posted March 18, 2018 Share #37 Posted March 18, 2018 test4321 Thanks for the advise......I have been looking at a couple of Supermicro server motherboards......still doing research... I have also read the SSD cache reviews...not sure applies to my current use...believe cache would be more useful for files that I would use often...according to the Synology website..." SSD cache can improve the performance of random access by storing frequently accessed data on the SSD cache. SSD cache can be mounted on a volume or iSCSI LUN (Block-Level)." https://www.synology.com/en-us/knowledgebase/DSM/help/DSM/StorageManager/genericssdcache ......currently my plan is for long term redundancy back ups....but I wonder if I decide to run virtual machines in xepnology would take advantage of this.... Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 18, 2018 Share #38 Posted March 18, 2018 (edited) On 1/31/2018 at 9:15 PM, mervincm said: I don't think that the mellanox connectX-2 (I tested one in the past) have any sort of built in driver support in XPenology. that is not the case, i already had the adapter ic listed so in theory it was possible to check but as it seems the connectx type name is always used so i added a listing of connectx types and ic familiy looks like every adapter from ConnectX (1) to ConnectX-6 is supported with the drivers that come from synology with dsm 6.1 for ds3615/3617 (916+ comes without this drivers from synology and my extra.lzma has only untested driver as there was no feedback about it, so it should be treated as may work or not - there is a special 916+ section to read about the shortcomings, 916+ is a consumer version and does not provide as much drivers as the business models) looks like cheap ConnectX-2/3 are not such a bad choice as they are natively supported by dsm as it comes from synology Edited March 19, 2018 by IG-88 1 Quote Link to comment Share on other sites More sharing options...
b4u Posted March 23, 2018 Share #39 Posted March 23, 2018 I have Hp microserver gen 8 running the latest DSM 6.1.5-15254. Booting as DS3617xs. I have installed BR-1020 10gbe network card. I am using jun's 1.02b loader with latest extra.lzma file (v4.6 11.03.2018) however it seems that dsm is not recognizing this card. I don't have 10gbe network interfaces shown in menu. Is there any setting in the menu to activate the interface? I am new to 10gb networking... Thanks Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 24, 2018 Share #40 Posted March 24, 2018 (edited) 5 hours ago, b4u said: I am using jun's 1.02b loader with latest extra.lzma file (v4.6 11.03.2018) however it seems that dsm is not recognizing this card. I don't have 10gbe network interfaces shown in menu if it works normal it should just show up, you gave no info from log but most common case would be it might need firmware so i checked the kernel source and found this /linux-3.10.x/drivers/net/ethernet/brocade/bna/cna.h ... #define CNA_FW_FILE_CT "ctfw-3.1.0.0.bin" #define CNA_FW_FILE_CT2 "ct2fw-3.1.0.0.bin" ... i added the files and created a new version 4.7 you can try -> https://xpenology.com/forum/topic/9508-driver-extension-jun-102bdsm61x-for-3615xs-3617xs-916/ Edited March 24, 2018 by IG-88 Quote Link to comment Share on other sites More sharing options...
RacerX Posted March 24, 2018 Share #41 Posted March 24, 2018 Small Test HP SSF INTEL Xeon D-1527 Installed DS3617 V1.02b for DSm 6.1 Added LSi HBA 9207 Mellanox MHQH29B-XTR ConnectX 2 The system is bare metal Results - The LSI HBA 9207 card is very transparent. It works fine right out of the box On the other hand the Mellanox MHQH29B-XTR ConnectX 2 does not show up under network interfaces, With SSH Me@Test:/$ lspci 0000:00:00.0 Class 0600: Device 8086:0c00 (rev 06) 0000:00:01.0 Class 0604: Device 8086:0c01 (rev 06) 0000:00:02.0 Class 0300: Device 8086:0412 (rev 06) 0000:00:03.0 Class 0403: Device 8086:0c0c (rev 06) 0000:00:14.0 Class 0c03: Device 8086:8c31 (rev 04) 0000:00:16.0 Class 0780: Device 8086:8c3a (rev 04) 0000:00:16.3 Class 0700: Device 8086:8c3d (rev 04) 0000:00:19.0 Class 0200: Device 8086:153a (rev 04) 0000:00:1a.0 Class 0c03: Device 8086:8c2d (rev 04) 0000:00:1b.0 Class 0403: Device 8086:8c20 (rev 04) 0000:00:1c.0 Class 0604: Device 8086:8c10 (rev d4) 0000:00:1c.4 Class 0604: Device 8086:8c18 (rev d4) 0000:00:1d.0 Class 0c03: Device 8086:8c26 (rev 04) 0000:00:1f.0 Class 0601: Device 8086:8c4e (rev 04) 0000:00:1f.2 Class 0106: Device 8086:8c02 (rev 04) 0000:00:1f.3 Class 0c05: Device 8086:8c22 (rev 04) 0000:01:00.0 Class 0107: Device 1000:0087 (rev 05) 0000:03:00.0 Class 0c06: Device 15b3:673c (rev b0) 0001:00:02.0 Class 0000: Device 8086:6f04 (rev ff) 0001:00:02.2 Class 0000: Device 8086:6f06 (rev ff) 0001:00:03.0 Class 0000: Device 8086:6f08 (rev ff) 0001:00:03.2 Class 0000: Device 8086:6f0a (rev ff) 0001:00:1f.0 Class 0000: Device 8086:8c54 (rev ff) 0001:00:1f.3 Class 0000: Device 8086:8c22 (rev ff) 0001:06:00.0 Class 0000: Device 1b4b:1475 (rev ff) 0001:08:00.0 Class 0000: Device 1b4b:9235 (rev ff) 0001:09:00.0 Class 0000: Device 8086:1533 (rev ff) 0001:0c:00.0 Class 0000: Device 8086:1533 (rev ff) 0001:0d:00.0 Class 0000: Device 8086:1533 (rev ff) Me@Test:/$ Not sure how to test it out any further, I only have this system for test this weekend then I have to give it back. Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 24, 2018 Share #42 Posted March 24, 2018 7 hours ago, RacerX said: 0000:03:00.0 Class 0c06: Device 15b3:673c (rev b0) just searched the web for the pci device vendor https://pci-ids.ucw.cz/read/PC/15b3 15b3:673c - MT26428 [ConnectX VPI PCIe 2.0 5GT/s - IB QDR / 10GigE] in kernel source of dsm 6.1 kernel /linux-3.10.x/drivers/net/ethernet/mellanox/mlx4/main.c static DEFINE_PCI_DEVICE_TABLE(mlx4_pci_table) = { ... { PCI_VDEVICE(MELLANOX, 0x673c), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT25408 "Hermon" EN 10GigE */ { PCI_VDEVICE(MELLANOX, 0x6368), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT25408 "Hermon" EN 10GigE PCIe gen2 */ { PCI_VDEVICE(MELLANOX, 0x6750), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT25458 ConnectX EN 10GBASE-T 10GigE */ { PCI_VDEVICE(MELLANOX, 0x6372), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT25458 ConnectX EN 10GBASE-T+Gen2 10GigE */ { PCI_VDEVICE(MELLANOX, 0x675a), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT26468 ConnectX EN 10GigE PCIe gen2*/ { PCI_VDEVICE(MELLANOX, 0x6764), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT26438 ConnectX EN 40GigE PCIe gen2 5GT/s */ { PCI_VDEVICE(MELLANOX, 0x6746), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT26478 ConnectX2 40GigE PCIe gen2 */ { PCI_VDEVICE(MELLANOX, 0x676e), MLX4_PCI_DEV_FORCE_SENSE_PORT }, /* MT25400 Family [ConnectX-2 Virtual Function] */ { PCI_VDEVICE(MELLANOX, 0x1002), MLX4_PCI_DEV_IS_VF }, /* MT27500 Family [ConnectX-3] */ { PCI_VDEVICE(MELLANOX, 0x1003), 0 }, /* MT27500 Family [ConnectX-3 Virtual Function] */ ... and synology seems to use even newer drivers (3.3.-1.0.4) as there is also a mlx5 module, which is not part of the original kernel, so you should work ootb i guess your dsm 6.1 is running with the card plugged in so have a look at /var/log/dmesg what it says about the card the driver is natively part of dsm and should load so there should be something in the log about it mellanox official supported cards for the 3.3-1.0.4 driver and needed min. firmware can be found here http://www.mellanox.com/page/mlnx_ofed_matrix?mtag=linux_sw_drivers Quote Link to comment Share on other sites More sharing options...
RacerX Posted March 24, 2018 Share #43 Posted March 24, 2018 It does.. dmesg -wH [+0.000002] Backport generated by backports.git v3.18.1-1-0-g5e9ec4c [ +0.007100] Compat-mlnx-ofed backport release: cd30181 [ +0.000002] Backport based on mlnx_ofed/mlnx_rdma.git cd30181 [ +0.000001] compat.git: mlnx_ofed/mlnx_rdma.git [ +0.053378] mlx4_core: Mellanox ConnectX core driver v3.3-1.0.4 (03 Jul 2016) [ +0.000008] mlx4_core: Initializing 0000:03:00.0 [ +0.000033] mlx4_core 0000:03:00.0: enabling device (0100 -> 0102) [ +0.420818] systemd-udevd[6199]: starting version 204 [ +1.251641] mlx4_core 0000:03:00.0: DMFS high rate mode not supported [ +0.006420] mlx4_core: device is working in RoCE mode: Roce V1 [ +0.000001] mlx4_core: gid_type 1 for UD QPs is not supported by the devicegid _type 0 was chosen instead [ +0.000001] mlx4_core: UD QP Gid type is: V1 [ +1.253954] mlx4_core 0000:03:00.0: PCIe BW is different than device's capability [ +0.000002] mlx4_core 0000:03:00.0: PCIe link speed is 5.0GT/s, device support s 5.0GT/s [ +0.000001] mlx4_core 0000:03:00.0: PCIe link width is x4, device supports x8 [ +0.000087] mlx4_core 0000:03:00.0: irq 52 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 53 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 54 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 55 for MSI/MSI-X [ +0.000004] mlx4_core 0000:03:00.0: irq 56 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 57 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 58 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 59 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 60 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 61 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 62 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 63 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 64 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 65 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 66 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 67 for MSI/MSI-X [ +0.000003] mlx4_core 0000:03:00.0: irq 68 for MSI/MSI-X [ +1.150446] mlx4_en: Mellanox ConnectX HCA Ethernet driver v3.3-1.0.4 (03 Jul 2016) Quote Link to comment Share on other sites More sharing options...
RacerX Posted March 24, 2018 Share #44 Posted March 24, 2018 Adjusted test, removed LSI 9207 and tested the ConnetX2 card in the first slot dmesg -wH [ +0.019285] Compat-mlnx-ofed backport release: cd30181 [ +0.000002] Backport based on mlnx_ofed/mlnx_rdma.git cd30181 [ +0.000001] compat.git: mlnx_ofed/mlnx_rdma.git [ +0.061974] mlx4_core: Mellanox ConnectX core driver v3.3-1.0.4 (03 Jul 2016) [ +0.000008] mlx4_core: Initializing 0000:01:00.0 [ +0.000031] mlx4_core 0000:01:00.0: enabling device (0100 -> 0102) [ +0.530407] systemd-udevd[5965]: starting version 204 [ +1.141257] mlx4_core 0000:01:00.0: DMFS high rate mode not supported [ +0.006462] mlx4_core: device is working in RoCE mode: Roce V1 [ +0.000001] mlx4_core: gid_type 1 for UD QPs is not supported by the devicegid _type 0 was chosen instead [ +0.000001] mlx4_core: UD QP Gid type is: V1 [ +0.750613] mlx4_core 0000:01:00.0: PCIe link speed is 5.0GT/s, device support s 5.0GT/s [ +0.000002] mlx4_core 0000:01:00.0: PCIe link width is x8, device supports x8 [ +0.000080] mlx4_core 0000:01:00.0: irq 44 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 45 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 46 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 47 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 48 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 49 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 50 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 51 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 52 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 53 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 54 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 55 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 56 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 57 for MSI/MSI-X [ +0.000002] mlx4_core 0000:01:00.0: irq 58 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 59 for MSI/MSI-X [ +0.000003] mlx4_core 0000:01:00.0: irq 60 for MSI/MSI-X [ +0.822443] mlx4_en: Mellanox ConnectX HCA Ethernet driver v3.3-1.0.4 (03 Jul 2016) [ +3.135645] bnx2x: QLogic 5771x/578xx 10/20-Gigabit Ethernet Driver bnx2x 1.71 3.00 ($DateTime: 2015/07/28 00:13:30 $) Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 24, 2018 Share #45 Posted March 24, 2018 looks like the driver is working did you change the nic settings in grub.cfg on your usb flash drive? like: set netif_num=3 set mac1=... set mac2=... set mac3=... asuming a one port nic (eth0) with a 2port mellanox whats in the log about ethX cat /var/log/dmesg | grep eth1 cat /var/log/dmesg | grep eth2 Quote Link to comment Share on other sites More sharing options...
RacerX Posted March 24, 2018 Share #46 Posted March 24, 2018 If it's working that is news to me. My test is stock DS3617xs 6.1 Jun's Mod V1.02b 7/4/2017 i did not change the usb stick. Do I need to change it for the test? The mellanox card is two ports. cat /var/log/dmesg | grep eth1 cat /var/log/dmesg | grep eth2 just returns nothing I connected the cable from one port to the other port since I do not have a 10GB Switch There are no link link lights and it does not show up with the network interfaces. Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 24, 2018 Share #47 Posted March 24, 2018 1 hour ago, RacerX said: If it's working that is news to me. in the way of it detects a hardware present and does not crash, so maybe something else is missing 1 hour ago, RacerX said: My test is stock DS3617xs 6.1 Jun's Mod V1.02b 7/4/2017 i did not change the usb stick. Do I need to change it for the test? i never tried what happens if you insert more nic's and do not change this settings, so yes try to change it, can be easily changed back later it you dont have the real mac addresses just make some up, its not important for testing it 1 hour ago, RacerX said: The mellanox card is two ports. i guess so, the model you gave was a 2port 1 hour ago, RacerX said: cat /var/log/dmesg | grep eth1 cat /var/log/dmesg | grep eth2 just returns nothing if they are not present then the gui will not show anything about more nic's Quote Link to comment Share on other sites More sharing options...
RacerX Posted March 24, 2018 Share #48 Posted March 24, 2018 (edited) Hi I have another single port card 02:00.0 Network controller: Mellanox Technologies MT27500 Family [ConnectX-3] In another computer if that I could test tomorrow that helps Edited March 24, 2018 by RacerX Quote Link to comment Share on other sites More sharing options...
IG-88 Posted March 25, 2018 Share #49 Posted March 25, 2018 (edited) only makes sense to test this if changing the grub.cfg fails, it its about the grub.cfg then the setting it for one card and it will make no difference if it is one or two more ports from the dmesg it look like the driver loads so the problem might not be the driver itself Edited March 25, 2018 by IG-88 Quote Link to comment Share on other sites More sharing options...
b4u Posted March 25, 2018 Share #50 Posted March 25, 2018 On 3/24/2018 at 2:38 AM, IG-88 said: i added the files and created a new version 4.7 you can try -> https://xpenology.com/forum/topic/9508-driver-extension-jun-102bdsm61x-for-3615xs-3617xs-916/ Thank you!!! I've updated extra.lzma and two more interfaces showed up in diskstation menu. Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.