• Content Count

  • Joined

  • Last visited

Community Reputation

0 Neutral

About unmesh

  • Rank
    Advanced Member

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. @IG-88, When I have some time to spare, I am going to try the method from this thread on a fresh test installation. Never know when I will do an update I shouldn't have attempted Thanks
  2. What has worked for me in the past is to disconnect the old drives, add a new (in my case small size) drive and flash a new bootloader USB. I then install the last known DSM version onto the new drive using the same login and password. Then I add back the original drives and DSM offers to correct the version on them (it might be the migrate option but I don't remember). I let it do that, power down, remove the new drive, let it reboot with only the original drives and all was fine. I keep good records of what is running and save the bootloader images with their modified grub.cfg an
  3. When I try creating VMs with arkilee's .vma file, I get the following messages during the restore though DSM does come up WARNING: You have not turned on protection against thin pools running out of space. WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full. What if anything do I need to do to fix this? Also, a display isn't configured but a serial port is. Using "qm terminal <vmid>" after starting the VM takes me some time. How do I open a terminal to look at the serial output
  4. Large file transfers between the VM on the Lenovo TS140 and the HP N54L have historically been limited to approximately 95MB/s by the Gigabit Ethernet link. I got 10Gigabit NICs for both and have recently got around to testing performance. For the network, iperf3 shows about 9Gbps transfer between the two. The ESXi server happens to have a NVMe drive in addition to hard drives so I used the former for one end of the storage transfer. The HP only has hard drives though it is a modern one and dd writes from /dev/zero occur at 250MB/s and dd reads to /dev/null occur at 150MB/s after f
  5. OP asked about putting the synoboot image in a datastore on a USB drive. I wanted to create USB drives with the latest version of 6.5, 6.7 and 7.0 to test which ones would boot on a variety of PCs and learned that one could create a datastore on the ESXi boot drive itself! The following link was instrumental. https://www.horizonbits.com/2017/02/19/squeezing-esxi-on-usb/ I haven't tried it as a datastore for deployment though.
  6. @bearcat, I could only boot with the USB stick with the wrong VID/PID and wanted to edit grub.cfg while running DSM. When I SSH'ed in, I noticed that there wasn't a /dev/synoboot1 available to mount. Having used FixSynoBoot.sh on ESXi, I remembered reading that it had helped with some baremetal installs so I tried it and it did! I will try and find some time to do a clean install
  7. It's the same USB stick, so by elimination, it's the image but I don't know how to test the image other than booting from it.
  8. I've had a N54L running for several years and thought getting a N40L running would be piece of cake. Something changed to make it difficult and I'm hoping to figure out what out of curiosity now more than anything else.
  9. The serial number is a previous serial number +1 and I don't remember how I got that one. The first MAC address is a Realtek based add-in NIC made by TP-Link while the second is the Broadcom embedded NIC I use Win32DiskImager too. And here is something new. I used Win32DiskImager to flash a USB disk with the image previously used for the N54L, booted the N40L with that stick and a new hard drive, installed DS3615xs 6.2.3, enabled SSH, installed FixSynoBoot.sh, mounted /dev/synoboot1, edited grub.cfg with the same information shown above, saved a copy of it t
  10. I've been running fine on a baremetal N54L and was able to get my hands on a N40L to use for remote backup. In any case, I added a PCIe NIC to it, mounted the saved synoboot.img file from my N54L installation into OSFmount and edited grub.cfg with a new serial number, the MAC addresses of the two NICs (I realize that is not necessary) and the new VID:PID of the boot USB stick. I flashed the .img file to the stick using Win32DiskImager and booted the N40 without any hard drives It seemed to boot but then failed with several error: can't find command 'common_add_option' messag
  11. I recently upgraded one of my GbE NICs on a N54L to a 10GbE NIC and was gratified to see that SMB Multichannel still worked with 10+1 GbE on the server and 1+1 GbE on my Windows 10 desktop. Throughput for large file transfers was understandably capped at 200-ish MBps. I then bought another 10GbE NIC to do a similar upgrade on my ESXi7 server which is configured with two vswitches, one for each of its GbE NICs. While ESXi recognizes the Mellanox NIC and I can log into the server on either of the two IP addresses, SMB Multichannel performance is only around 100-ish MBps.
  12. I found a post that talks about compiling a driver for the RTL8125 for the 1.04b DS918+ loader.
  13. Very nice form factor! Though dual 2.5Gbps is definitely overkill for the application Does DSM have drivers for the RTL8125? Thanks.
  14. I'd like to put together a x86/x64 1-bay NAS with a 10TB 3.5" hard drive and drop it off at a friend's location as a remote NAS backup target but the SFF/Thin-Client PCs I see are typically designed for 2.5" drives. Maybe I haven't looked hard enough and someone can suggest devices. An alternative would be to hang a USB 3.0 drive off a power sipping thin client running Xpenology which would look a little kudgey. I've found the instructions to have DSM use USB drives as internal drives but is this a bad idea? Thanks
  15. My (failing) memory though that was what I had done originally and I managed to find some notes that said I should use the first entry so you are correct. I also decide to compare all the BIOS settings against my notes and discovered that C1E had somehow gotten enabled . Weak CMOS battery, perhaps? I suspect that putting back the old USB stick and hard drives will now work. The Mellanox is first in PCIe enumeration so the only way to have a 1Gbe first is to take the card out and optionally put the Intel NIC back. Good news is that the system booted and I could access th