Jump to content
XPEnology Community

JBark

Member
  • Posts

    16
  • Joined

  • Last visited

  • Days Won

    3

Everything posted by JBark

  1. - Outcome of the update: SUCCESSFUL - DSM version prior update: DSM 6.1.7-15284 Update 2 - Loader version and model: JUN'S LOADER v1.03b - DS3617xs - Using custom extra.lzma: NO - Installation type: BAREMETAL - HP MicroServer Gen8 with XEON 1265v2 - Additional comments: Previous attempts to upgrade from 6.1 to 6.2 failed, as all services, including networking, would shutdown a few seconds after logging in to DSM. Saw other posts mentioning an old .xpenoboot folder in / as the cause, so I deleted that and upgrade worked fine. I've been running XPenology on this same server for over 3 years now, so I'm not surprised there's a bunch of old crap laying around that could cause these problems.
  2. JBark

    DSM 6.2 Loader

    That's interesting, because I've also added that line to my synoinfo.conf file in 6.1. I'd completely forgotten about it, so I'll mess around with 6.2 again and see if this might be the trigger. Those scripts in /etc are suspicious as well, because they don't exist in 6.x, but it sure looks like /etc/upgrade.sh is being called on my system, when I watch 1st boot output using VSP through iLO. I wonder if they're some holdover from 5.x and are accidentally being called during this upgrade, and messing everything up. Oh well, it's a place to start.
  3. JBark

    DSM 6.2 Loader

    HP MicroServer Gen8 with E3-1256L V2 boots fine with new loader and upgraded OK from 6.1 to 6.2, but getting some odd behaviour when I log into DSM. If I don't mess with DSM, it runs fine, including the VM I have configured under VMM. The second I log into DSM, it shuts down every service and I lose connection about 30 seconds later. It's not a hard crash, as I can watch the logs and see it's shutting down/pausing every service that's running. Pretty odd, I'm guessing something didn't upgrade nicely from 6.1 -> 6.2, I doubt it's loader related. From /var/log/messages: 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgtool.cpp:2284 chmod failed for /var/packages (Operation not permitted) 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgtool.cpp:2293 chmod failed for /usr/local/etc/rc.d (Operation not permitted) 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-phpMyAdmin (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-PHP5.6 (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-MariaDB10 (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-Apache2.2 (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-FileStation (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-TextEditor (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-tvheadend (err=-1) [0x0900 servicecfg_root_object_set.c:43] 2018-08-02T12:20:52+08:00 HPGEN8 synopkg: pkgstartstop.cpp:379 Failed to stop pkgctl-SynoFinder (err=-1) [0x0900 servicecfg_root_object_set.c:43] Tons of entries like that, followed by.... 2018-08-02T12:21:55+08:00 HPGEN8 ore.Desktop.Initdata_1_get: service_pause_all.c:173 Pause service [synosnmpcd] by reason of [unidentified] failed. 2018-08-02T12:21:55+08:00 HPGEN8 ore.Desktop.Initdata_1_get: service_pause_all.c:173 Pause service [synogpoclient] by reason of [unidentified] failed. 2018-08-02T12:21:55+08:00 HPGEN8 ore.Desktop.Initdata_1_get: service_pause_all.c:173 Pause service [crond] by reason of [unidentified] failed. 2018-08-02T12:21:55+08:00 HPGEN8 ore.Desktop.Initdata_1_get: service_pause_all.c:173 Pause service [iscsitrg] by reason of [unidentified] failed. 2018-08-02T12:21:55+08:00 HPGEN8 ore.Desktop.Initdata_1_get: service_pause_all.c:173 Pause service [atalk] by reason of [unidentified] failed. Continues like that until I lose the connection. If I check messages after the next boot, it's shutdown everything like it's planning to reboot or something, but never actually finishes.
  4. Interesting thing to note is that Ports 1&2 are SATA3, Ports 3&2 are SATA2. Googling around shows lots of other hits from people having issues with these new Gold drives not being detected, and in one case somebody reported there were only working at SATA2 speeds. I'd say there's probably a firmware bug in the drives that's affecting speed negotiation with the chipset at boot time, and the firmware that's loaded in AHCI vs RAID mode is different enough that it works. Some WD SATA drives have jumpers where you can force the speeds of the drive back to SATA2, but doesn't look like this drive has any jumpers on it.
  5. Yeah, as far as I can tell, it requires the 2nd interface, since it binds one NIC for management traffic (clustering), and the other NICs are for VM traffic. I don't think there's any reason why they couldn't do it all on 1 NIC, but requiring at least two means they can guarantee that VM traffic won't affect cluster traffic (and access to DSM itself), and vice versa. Would be kinda nice if they had an option to disable clustering, so you could run in single host mode, but I think this is designed mainly for the SMB side of things, and they probably want to get rid of as many points of failure as possible.
  6. I just switched to using the new Synology Virtual Machine Manager that's been in beta for a couple weeks: https://www.synology.com/en-global/beta ... ualization It's not anywhere near as full featured as phpVirtualBox yet, but it's the usual qemu and libvirt that is common on LInux, and does run my Windows VM just fine. I'll never have to worry about compatibility issues again when upgrading DSM, which is great. I would not recommend trying to migrate a VM from phpVirtualBox, to VMM, it was an absolutely nightmare. There's also some gotchas, you must have 2 or more LAN connections, and you can only put the VMs on BTRFS volumes. No workaround for either that I've seen. The problems I ran into were twofold. One, VMM seems to not support UEFI yet, and I had my vbox vm using UEFI and GPT partitions. Two, and this is the big problem, VMM only supports VirtIO SCSI for the drives, and it's virtually impossible to install these drivers cleanly in a pre-existing VM. If they were using the regular qemu setup, you could migrate over a VM, set the controller to IDE, and boot up fine the first time. Add a 2nd small drive using VirtIO SCSI to get Windows to properly install the drivers, then shutdown, remove the 2nd drive, and change the primary over the VirtIO SCSI. However, this isn't possible with VMM. So here's what I did. First, I copied over the vdi and vbox files from the NAS so I could fix everything up on my local PC, since it's much faster. UEFI to MBR: I imported into VirtualBox, added a 2nd drive the same size as the primary, and booted up with a GParted live iso. Initialized 2nd drive using MBR, then copied the partition from the GPT to MBR drive using GParted. Shutdown VM, replaced GParted ISO with Win10 ISO, booted back up, ran Command Prompt from Recovery menu ran these commands: bootrec /scanos bootrec /rebuildbcd bootrec /fixmbr bootrec /fixboot Rebooted back into recovery again, ran Startup Repair, booted back into Windows without issue after that. Getting VirtIO Drivers installed: This is annoying, because even though you can right-click install the inf files, Windows won't complete the install until it sees the hardware, and this can't be on first boot. So you somehow need Windows to "see" the VirtIO hardware and install the drivers before you can switch the boot drive. And VirtualBox doesn't support VirtIO SCSI (just VirtIO network), so much more work is needed. Here's my painful steps. QEMU works much better under linux, so I created an Ubuntu VM with VMWare Player (vbox can't do nested virtualization, player can if you enable it), copied my VDI file from vbox to it, and installed the qemu files sudo apt-get install qemu-kvm libvirt-bin virt-manager (something like that). Converted the VDI to a format qemu likes qemu-img convert -p -f vdi vbox.vdi -O qcow2 qemu.qcow2 Created a VM using virt-manager that matched the specs of my VM, added the new qcow2 disk image as an IDE drive. Fight with VM settings to actually get something to boot. Booted VM up, confirmed working. Added a 2nd 1GB drive using VirtIO SCSI, windows saw device, installed drivers from VirtIO iso: https://fedorapeople.org/groups/virt/vi ... io-win.iso Added a few other drives using the other VirtIO types, since I didn't want to mess with this again. Powered down VM Important step, have to remove the VM (leaving the qcow2 image along), then setup a new VM again, this time chosing to add the qcow2 drive as VirtIO. When I just changed the drive from IDE to VirtIO, it wouldn't boot, and googling found I had to remove and reimport. Dumb. Powered up VM, all good. Powered down, converted back to a format vbox can use (I used vmdk cause I need to export to OVA anyway) qemu-img convert -p -f qcow2 qemu.qcow2 -O vmdk vbox-fixed.vmdk Copy vbox-fixed.vmdk back to PC with vbox, replace old vdi in VM with this new vmdk Export the vm to an ova file. Uninstall phpVirtualbox, upgrade NAS to 6.1.1 (required for VMM), enable beta software, install VMM, configure VMM, import ova file, cross fingers. Worked for me, though still some weirdness that the VM gets stuck in a boot loop if I reboot it, but is fine if I shutdown/power up. So it looks like things are borked as Windows does the first boot and installed devices and reboots, but just needs a forced power off/back on to be fine. It took me literally days to get it working, and it was totally not worth it. Could have easily just created a new VM and reconfigured it in far lass time, but I was mainly curious to see if it was actually possible. If you're just starting out, and you meet the 2+ LAN and BTRFS requirements, definitely look at VMM. It's getting a bunch of updates to add new features roughly every week, so should be quite nice in a couple months.
  7. This latest beta version has been working fine for me on my HP Gen 8 running Jun's 1.02 Alpha loader and 6.1 Update 1. I did notice it didn't automatically shut down the guest VM when I rebooted DSM, but I'm not sure if that has ever worked.
  8. JBark

    DSM 6.1.x Loader

    Yep, running Update 1 without any problems, other than the usual false alarm "storage space has failed" message I get at boot.
  9. JBark

    DSM 6.1.x Loader

    Confirming this worked for me as well on my Gen8. I needed to get up to 6.1 so I could finally update to the newest phpVirtualBox, and this was my only option. I'm getting the "failed storage space" for my single SSD at boot, but everything runs fine so that's not really a big deal.
  10. Thanks guys for putting out an updated version for DSM 6. I've been meaning to get around to it, but haven't bothered updating since phpVirtualBox hasn't been updated yet. I'll update the OP here sometime soon with the links to this new version.
  11. I can confirm that the -5 version has fixed the bridging issues I was seeing as well with the latest bootloader. Nice to be able to run the latest everything again without having to resort to any sort of workaround.
  12. It does work if you can downgrade xpenoboot to the previous version. After updating dsm to 5.2 I flashed xpenoboot back to 5.2-5592.2. Everything has been fine for the past month or so, but it definitely doesn't work on every setup, so I can't say it's a guaranteed fix. Once we get a fixed xpenoboot then it should work again.
  13. On the HP Gen8, rebooting DSM will cause the BIOS to reset to defaults. It's annoying, but you can workaround it by configuring everything correctly in the BIOS, then saving those settings as the default. Then when it resets to default, it will just reset to those settings. AHCI is the preferred way to go, since SMART values on the drives can't be read in RAID mode. There used to be some issues with fan speeds increasing when running in AHCI, but that should be fixed if you've got the latest firmware installed.
  14. JBark

    Baremetal vs ESXi

    If you're running bare metal, you can use this VirtualBox package to run VMs. viewtopic.php?f=15&t=3497 Just scroll to the end to find the post with the most recent package. Works great, I've been using it for the past couple months to run a Windows VM without issue.
  15. I'm running bare metal, with the B120i in AHCI mode. The fan speed depends on the intake temp, so I probably just have slightly lower temps that you. I just checked today, the intake temp is 21C and the fan is at 11%. I think AHCI is always slightly higher than RAID mode, though it used to be much worse with the older firmware, where the fan was around 30% in AHCI mode. There are a couple other hp apps that get installed on Linux, like the hp-health one. It could be that one of those are needed as well to get the same fan speeds we see in supported OSes? I'm not really sure.
  16. I've spent the past couple days messing with this, and I've mostly got it working, though there are still a few bugs. The HP AMS code assumes a lot of things exist on the system that obviously aren't there with DSM, and that means it segfaults non stop. I finally got a build to work, and the main bug I notice is that in iLO only the first HDD is listed. This is in AHCI mode, it's possible the drive details show up OK in RAID mode. Also, I didn't really see any fan speed changes once I got this working, it was 10% before, and 9% after. Here's a dump of what I've been testing. No idea how correct any of it is, haven't done any development in a long time This was built against the latest bromolow DSM source and is running on 5.2-5592 Update 3, I dunno if it works on other versions of DSM. Use at you own risk, though I can't imagine there's much risk involved with something that is basically just an SNMP wrapper. Kernel Module - Steps mostly from here: http://xpenology.com/wiki/en/building_xpenology Download DSM 5.2 source and toolchain: http://sourceforge.net/projects/dsgpl/f ... z/download http://sourceforge.net/projects/dsgpl/f ... z/download Extract toolchain to /usr/local, extract source somewhere cd source/linux-3.10.x cp synoconfigs/bromolow .config make ARCH=x86_64 CROSS_COMPILE=/usr/local/x86_64-linux-gnu/bin/x86_64-linux-gnu- menuconfig Go into Device Drivers -> Misc Devices, hit M at "Channel interface driver for the HP iLO processor" to build the kernel module Save and Exit make ARCH=x86_64 CROSS_COMPILE=/usr/local/x86_64-linux-gnu/bin/x86_64-linux-gnu- modules copy drivers/misc/hpilo.ko to /usr/lib/modules/3.10.35 on NAS vi /usr/lib/modules/3.10.35/modules.dep and add this line: /lib/modules/3.10.35/hpilo.ko: modprobe hpilo should work with no errors lsmod | grep hpilo should show that hpilo is loaded HP AMS I couldn't get any of the executables provided in the debs or rpms to run without segfaulting. So I searched around, and found some source on github: https://github.com/marker55/hp-ams It's not the absolutely latest from what I can tell (seems to be 2.1), but it's close, as 2.2 is the most recent rpm I could find. What's important is it has the new command line args that 2.0 and below don't have, which let use chose which MIBs to query So git clone that repo, or just download the source zip. Download http://sourceforge.net/projects/net-snm ... z/download and copy it to the hp-ams-master folder Run make From what I can tell, it's not too important what compilers are used to build. I wasted a day adding cross compile support into all the Makefiles so I could build with the Synology toolchain to see if that would fix the numerous segfaults, but it doesn't seem to be necessary. After build is complete, copy the hpHelper executable to /sbin/hpHelper on the NAS chmod +x /sbin/hpHelper Test it Assuming you've already modprobed hpilo, run this as root to test if AMS works /sbin/hpHelper -f -L -I0 -M1,5,14,16,18,23,99 Should see a message like 'NET-SNMP version 5.7.2 AgentX subagent connected' Give it a minute or so, and see if things start showing up in iLO. There are two other MIBs, 0 and 11, but including either of those just causes it to segfault. Might be fixable, but I haven't dug into the code. Run at start Ctrl-C your test run to stop it, then create a startup script. vi /usr/local/etc/rc.d/S99HP-AMS.sh #!/bin/sh modprobe hpilo /sbin/hpHelper -I0 -M1,5,14,16,18,23,99 chmod +x /usr/local/etc/rc.d/S99HP-AMS.sh /usr/local/etc/rc.d/S99HP-AMS.sh ps | grep hpHelper Should show it is running, and iLO should still report correctly. Download 5.2 - hp-ams.tar.gz 6.0 (thanks to ema) - https://mega.nz/#!gt0HTCJZ!PeXq-qutvb6mZT7ZuP3Htfi-abkEdOYNZ_HcCVTd-6Y 6.1 (thanks to ema) - https://mega.nz/#!R8dUhDJb!dyYEVSsrWSohPfU5O7Uxo-GisuMKZPZbG4HbeiIeHk0 Note the modules.dep.hpilo file, that's so you don't extract blindly and overwrite say the modules.dep file for VirtualBox. Rename to modules.dep if you don't have one, or copy the contents into your existing modules.dep. Unfortunately, I really don't have a lot of time to work on this, so I doubt I'll ever have a proper package available or keep things up to date. But I figured I could at least dump what I had, so anyone else that wants to mess around with it can.
×
×
  • Create New...