Share via

Hyper-V host VMs unable to communicate with VMs on other clustered host?

Jon Hall 5 Reputation points
2025-07-31T21:06:44.3333333+00:00

Hey all - I'm hitting a wall with this one and was hoping someone might spot something obvious I'm missing. Sorry if my description gets a bit dizzying...

I have 2 Server 2022 hosts in a failover cluster, each with similar networking:

  • 2x 40Gbps links for guest VM traffic (1 to each of 2 Aruba AOS-CX switches with MC-LAG configured between the switches, though no LAG/LACP configured on these particular VM interfaces), configured in single virtual switch with SET (port-based balancing).
  • 4x 25Gbps links to storage, via 2 separate storage VLANs via 2 independent switches
  • 2x 1Gbps links (OS-teamed) for management traffic
  • 2x 1Gbps links (OS-teamed) for migration traffic
  • Each host has a unique range for assignable hardware/MAC addresses, and I've even gone so far in troubleshooting this issue as to assign unique static MACs to each VM (did not resolve issue).

Hosts, switches, and clients elsewhere in the network all have no issues communicating with each other. The issue I find is with traffic specifically from guest VMs on one host to the other.

Host A's VMs have zero issues whatsoever pinging/communicating with VMs on Host B. They also have no issues communicating with anything else on the network.

Host B's VMs can't seem to ping Host A's VMs, UNLESS I connect to the destination VM and start a ping back... then suddenly the B VM's pings will begin getting responses, and the host B VM then has no issues pinging the Host A VM until it restarts... then the problem resumes.

Even if I ping from the Host A-side VM so that the B VM will be able to start getting replies to its own ping requests, it continues to have issues pinging any other VM on the host A-side unless I repeat the same process from every host A-side VM (with a couple dozen VMs, this is obviously a pain to deal with).

The other solution I've found is that if I live migrate the VM from Host B to Host A, then it has no problem communicating with the other existing VMs on Host A, and will continue to be able to ping them if I live migrate the VM back to host B. It will continue to be able to communicate with them until it is restarted... then the problem comes back.

I've set up Wireshark to monitor both the interfaces on the Aruba switches, as well as on the Host B server itself monitoring the interfaces in the virtual switch team - when I start a ping from a host B VM to a host A one, none of the wireshark captures detect any of the ping traffic whatsoever... that is, UNTIL I connect to the destination VM on the Host A side and start a ping back to the Host B VM... then the Host B VM starts getting its replies, and all ping traffic both ways instantly begins appearing in the Wireshark captures.

At no time do the host B VMs have any issue pinging any other devices on our network, nor does it have any issue communicating with Host A's mgmt address... they only have issues with Host A-hosted VMs.

I've verified all VLANs and virtual switch settings are consistent across the switches and both hosts' configurations.

Is there anything obvious I'm overlooking that might explain this sort of behavior? Any suggestions?

Thank you!

Windows for business | Windows Server | Storage high availability | Virtualization and Hyper-V
0 comments No comments
{count} vote

7 answers

Sort by: Most helpful
  1. Łukasz Duczmal 16 Reputation points
    2026-02-12T12:07:02.44+00:00

    Hi

    I had similar problem with communication.

    My environment:

    • Hyper-V cluster on 2 HPE ProLiant DL360 Gen10 servers.
    • OS: Windows Server 2022 Datacenter.
    • virtual switches:
      • ComputeSwitch - SET - uplinks: HPE Ethernet 10Gb 562SFP+ (driver version: 1.22.88.0)
      • ManagementSwitch - SET - uplinks: HPE Ethernet 10Gb 2-port 562FLR-T (driver version: 1.22.88.0)
    • Top of Rack switches: Juniper

    My observations:

    1. Pinging of VM11 located on NodeA from VM21 located on NodeB - it works if VM21 has MAC address of VM11 in ARP cache
    2. Pinging of VM11 located on NodeA from VM21 located on NodeB - it DOES NOT work if VM21 does not know MAC address of VM11. I observed that "Who has?" ARP packet is sent by VM21 to broadcast (FF-FF-FF-FF-FF-FF). This packet I saw on vmswitch of NodeB and ingress and egress ports of physical switch, but not on vmswitch of NodeA
    3. Pinging of VM11 located on NodeA from VM21 located on NodeB - when I changed MAC address of pinging vm (VM21) to something other than MAC range offered by Hyper-V (I set 76:4a:17:d2:33:05), then pinging VM11 from VM21 does work even then VM21 does not know MAC address of VM11. I observed that in this case "Who has?" ARP packet is visible on vmswitch of NodeA

    But to change MAC address on vm one has to power off this vm. I had some VMs which I couldn't power off even for a moment, so I decided to downgrade netcard drivers - it also solved communication problem.

    I think that observation no 3 is very interesting

    Best regards

    Łukasz Duczmal

    0 comments No comments

  2. Martijn Goorman 5 Reputation points
    2025-08-14T20:02:27.5166667+00:00

    My issue is fixed with a downgrade of the NIC driver!
    See: https://community.hpe.com/t5/proliant-servers-ml-dl-sl/hyper-v-network-issue-with-hpe-10gb-2-port-562flr-sfp-x710/td-p/7253209

    This driver supports many Intel X710 10/25/40GB based NIC's.

    There should be an Intel driver also, but unknown for me what version that should be.

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.