• Content Count

  • Joined

  • Last visited

  • Days Won


benok last won the day on November 22 2018

benok had the most liked content!

Community Reputation

2 Neutral

About benok

  • Rank
    Junior Member

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. benok

    8th gen i5 + Asrock H370M-ITX/ac works w/ DSM

    If you want to accelerate copying/moving files via SMB, You can try SMB multi channel, by editing smb.conf manually. I don't have experience, but some Synology users got doubled or 4 times transfer speed, gathering 2,4 links with this. SMB multi channel working in 6.1 release - Synology Forum SMB 3 Multichannel support on DSM 6.2? : synology SMB Multichannel: How It Works & Troubleshooting Guide | Level One Techs Let us know the result, if you try this : -)
  2. Hi, Thank you for posting this very very useful matrix. I noticed that working NIC for esxi is not e1000, but e1000e for me. Could you confirm ? (kernel panic log from serial console)
  3. benok

    DSM 6.2 Loader

    Do we need to add "disconnect all display cables from video cards" to the check list ? I've read that on above post, but recently no one wrote about display cable. Is that still required to work transcoding ? I didn't have experience those because I don't have 918p enabled hardware, but I guess some of you might miss display cable disconnection.
  4. benok

    DSM 6.2 Loader

    @Tattoofreak, did you confirm CPU options of your DSM VM on ESXi ? I think you have to set options below if you didn't yet. Enable "Expose hardware assisted virtualization to the guest OS" Choose "CPU/MMU Virtualization" to "Hardware CPU and MMU"(or try another) It's necessary to enable those to work nested virtualization, generally. # But I'm not confident because VDSM error message didn't point out virtualization capability. # So, I think you did those already...
  5. benok

    NVMe optimization & baremetal to ESXi report

    Ah really, so that screenshot says it's "Cache Device". I didn't notice such limitation exists in NVMe on DS918. If NVMe is brought to higher models like DS3619(?), they would remove such a limit. I have no experience of cache corruption, using SSD cache several years, but it's heavy access wares SSD more, and SSD cache doesn't accelerate sequential read/write. Umm, I was attracted huge datastore & good random access performance for < ~10 VMs online. I agree that is not efficient resource usage. (But it's not so bad job for old SATA SSD.) I also use SSD only datastore (via DSM nfs) on another env, but I should consider my setup again. (I just migrating data & app from slow DSM5.2 baremetal NAS to my esxi server with new external enclosure.) As far as my memory is correct, several business vSphere users heavily complain about bad performance of btrfs on DSM6 beta (& Syno's support). (but I'm not sure it was nfs or iSCSI.) That thread continues after DSM6 release. So I thought I should not choose btrfs for VM datastore. I googled Syno's forum again, but no luck. (It might be deleted by Syno because of such a content and it was beta thread.)
  6. benok

    NVMe optimization & baremetal to ESXi report

    Hi, again. (Your post is very interesting for me. I think I should follow you. ) Have you ever tried NVMe passthrough with 918+ VM ? As you know, some 918+ baremetal users successfully setup NVMe as a SSD cache. I wanted to know is it possible to work 918+VM with NVMe pass through like other OS's VM. I think pRDM is a bit more cumbersome than passthrough because of vmdk setup. (But I appreciate your sharing experience about those pRDM. ) p.s. My configuration is something like below. I've read btrfs performance is very bad for VM workload on some Synology forum. So I choose btrfs VM over ext4 fs. I don't measure performance seriously, but I'm satisfied using that. With this setup, you can safely upgrade child DSM VM using snapshot, so it's easier to follow new DSM. It's not recommended for everyone, but for those who uses many VMs or who loves flexibility than simplicity. (But I'm thinking recently this design is not so good because of inefficient energy usage, complexity, can't directly migrate to Synology...) p.s. 2 It's related only to esxi, but It might be possible to passthrough iGPU for transcoding to child DSM VM . (with some recent XEON, or Core i) I don't have hardware(& budget ) to try NVMe, iGPU passthrough now, but I want to know if is there someone who has tried those. Sorry for a bit thread hijacking ..
  7. benok

    DSM 6.2 Loader

    You can use serial console. EliteDesk has a serial port. Buy a USB serial cable and connect HP and another PC with that. You can see some boot log and can use serial console using some serial terminal app like Putty. I guess, if you disable serial port on BIOS, that might stop booting. I've once experienced that on a VM without serial port.
  8. Wow, thank you for sharing all of this. I don't have enough time to try this, but I want to someday soon. It's quite interesting and useful. I think it's better to post this contents to another thread to notice at least DVR users. Currently, I feel I have to check every contents in this forum, because of the improper or inaccurate thread title.
  9. Very interesting. NanoPi Neo behaves something like virtual USB (network) mass storage. right ? I think it's awesome idea. What device you connect this to ? Personally, I don't have any good usage for this now, but someone using DVR function of TV or gamer using PS4, XBox, etc. may be interested in this. Could you share more detail on somewhere ?
  10. SHR has several drawbacks on upgrade, if you want to use SHR, I recommend you to read this article carefully. Don't Roll Into Trouble When Expanding NAS Storage - SmallNetBuilder I stopped using SHR after I read this.
  11. I've read that boot loader of DSM6.x is required to be stored on read-write file system. Is it possible to make ISO bootable boot loader using cloop, tmpfs, overlay-fs as Live CD does ? (-> Building Your Own Live CD | Linux Journal ) If it's possible, we can run DSM6 on VPS hosting server. On VPS hosting environment, we can't easily add virtual disk. (I think if we want to run DSM6, we have to use qemu.) Is there any information required to persist on "boot loader partition" on installation process ? # Even so, we can replace iso image after installation. (copy back from vmdk of local vm install) If so, only we have to do is just cheat the installer to believe it's writable file system, is it ? @jun or other wizards here, how do you think ? Is it difficult to make ? # I think it's too simple, so there must be some pitfalls
  12. benok

    Loader and RDM

    I don't know how to support that, but I think it's hard to analyze & fix DSM & keep update and so on, and I think it have very little merit. I don't recommend to use RDM for XPEnology. I think RDM's merit is only for Vt-d unsupported CPU system can get some information from disk. It's not useful to configure & uneasy to replace disk, etc. You can't replace disk on crash, before you write rdm config for the new disk. I only used RDM on very old system which doesn't support Vt-d. I recommend to use pass-through SATA I/O chip (onboard / PCIEx) and choose vt-d enabled CPU, if you want to use esxi. My choice of setup for esxi system is something like this. It's for home workstation/server setup & not only for NAS, so I know it's not for everyone...
  13. I don't know such limitation exists. Is there before 5.2 ? I used 240gb x2 SSD for read/write cache for 3TBx4 RAID5 for around 2 years ? since DSM5.2. Last year I upgraded to another system with 1TB x2 cache and it also works fine. I think you should check the recent document again. It almost only says about the memory requirements. DSM caches frequently used random access area automatically within the cache size. (I also recommend not to enable sequential cache, as document says. It just works fast only for a while.) Optimal cache size depends on your workload. I'm using 10-15 VMs constantly, but my cache usage is just 45%. (Before upgrade, cache usage was very high. I can't recall the real number.) My SSD cache seems to be over spec for my current workload. If we can use Virtual SSD for SSD cache, we could share SSD for cache and datastore, but I think it's not so good both for complexity & performance loss. I didn't have investigated, but I guess virtual SSD don't have required command for SSD cache or returns bad response for some command, and DSM refused VSSD. You should log in to console via ssh and check logs around enabling SSD cache. We might get work with flag tweak around VSSD, if such flags exists... I don't have good idea for small factor server like your HP Gen8. I built my system with mid-tower PC case as I wanted to use it as both for workstation & server.
  14. As far as I know, virtual SSD can't be used as SSD cache. In my experience, SSD cache can only be used with pass-through-ing host SATA I/F and SSDs. If you tried with success, please post about that. BTW, My recommended configuration for ESXi system is following. In the following config, whole system is hosted with SSD cache & (almost) all drives are managed by DSM. It performs well and notifies me on any disk troubles. I'm satisfied with this configuration. (I think) It's not so complex, but has enough performance & good flexibility. I hope this helps. My recommended XPEnology based configuration of ESXi system: boot ESXi from USB drive (as you do) Add 1 disk for VMFS datastore (& use it's disk inteface directly for ESXi). (this datastore is just for booting "Host DSM" VM) (*1) Make 1 XPEnology VM as "Host DSM" VM and pass through All disk interfaces (other than above one) to the VM. (This VM is used only for ESXi datastore & ESXi host.) Add all other HDDs / SSDs to make XPEnology VM & setup SSD cache & format disk group with ext4 (for VM performance). Create share folder for ESXi datastore using nfs export (better for performance & good maintainability with SMB access. you can add SMB access for direct maintenance of the datastore from client PCs.) Add that nfs exported datastore from ESXi Add your own VMs on that datastore (*2) Add Another XPEnology VM ("User DSM" VM) with thick provisioned vmdk, formatting with btrfs (for usual file sharing, etc.). Add users & apps only on "User DSM". (*3) *1) If you don't use USB sharing for VM, you can use USB disk for this datastore, perhaps. *2) I can also add Windows/MacOS Desktop VM with pass-through-ing GPU and USB. (Choose ESXi 6.0 for hosting mac. You can still use vCenter 6.5 or later.) *3) The only I wish but I can't is H/W encoding with DSM6.1 + DS916 VM. (Perhaps, you have to pass-through host's iGPU. I don't have such iGPU system.)
  15. benok

    5.2 NFS VAAI broken?

    Recently I migrated my system from DSM5.2 to DSM 6 with NewLoader and checked VAAI feature. It works fine with NFS again. I can use thick provisioning with NFS and offline cloning is also working. There's no problem with SSD cache. I'm very happy with DSM6 with NewLoader. p.s. I didn't try to use btrfs yet for datastore, because I've heard btrfs has very bad performance as Virtual Machine's datastore... ZFS, BTRFS, XFS, EXT4 and LVM with KVM – A Storage Performance Comparison (2015) | Hacker News ZFS, BTRFS, XFS, EXT4 and LVM with KVM – a storage performance comparison ... print=true I'm interested in Snapshot Replication, but it's very hard to convert big datastore & check performance for me because of the lack of available hardware resources. ... ection_mgr If someone tested performance of btrfs, please share the result on the forum