Quantcast
Channel: Hyper-V forum
Viewing all articles
Browse latest Browse all 19461

Hyper-V 2019 2 node cluster - intermittent host / cluster issue when restarting a VM

$
0
0

Hi all,

Hoping someone may have ideas on how to resolve this one!  We have a 2-node cluster (HPE ProLiant DL360 Gen10) built on Windows 2019 Standard GUI, iSCSI to Nimble storage.  All validation tests pass with flying colors!  When we were testing, i.e. with very little load we had one instance of a 2019 VM (vm version 9) which got stuck in a 'stopping-critical' state when rebooting.  At this point, the host has gone 'awol' with cluster functions, from then on any cluster operation failed until the host is rebooted (draining would never finish, any other live migration times out and fails).  We found that we had a switch down so put it down to that (even though redundant paths were AOK).  Since then it has been working beautifully as expected, I've done plenty of migrations, draining of hosts without issue, until yesterday.  The client decided to patch a Window 7 VM, it rebooted and got stuck in the same fashion, Stopping-Critical.  I have tried the trick of killing the VM process, this does not work in this scenario.  There is nothing we can do expect cold boot the server, forcing failover to the other host.  Obviously a bit of an issue as the 30 servers now running on there restart on the other host, interrupting file / print services etc.

I've read about disabling VMQ but I really don't want to go disabling the advanced performance features unless we absolutely have to.  I am using Dynamic memory on most of the machines, otherwise pretty standard configurations.

No issues with resource we are using under half memory on both hosts.

Any ideas?

Thanks,
Simon


Viewing all articles
Browse latest Browse all 19461

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>