Hello guys,
maybe you can help me with following issue. I have 
created a little cloud with a host and two worker nodes using opennebula. The 
setup went 
successfully until now, I am able to create VM's and move them via 
normal and live migration.
Another (possibly) important information is that I configured my virtual bridge 
on both worker nodes like this:


auto br0 iface br0 inet static address 192.168.0.[2|3] netmask 255.255.255.0 
network 192.168.0.0 broadcast 192.168.0.255 #gateway 192.168.0.1 bridge_ports 
eth0 bridge_stp on bridge_maxwait 0
The command "brctl show" gives me following things back:


bridge name    bridge id            STP enabled    interfaces
br0            8000.003005c34278    yes            eth0
                                                   vnet0 (<- only appears on 
node with running VM)

virbr0         8000.000000000000    yes

According to thelibvirt wiki this setting is good as is. However, the issue I'm 
having is that when I create a VM and assign a static IP to it, which looks 
like e.g. 192.168.0.5, 
I firstly am able to ping this VM from both worker nodes, and also when I 
perform a live migration the ping stops for a few seconds (until the nodes 
realize the new route to this VM) and then 
starts pinging normally again.

However, when I perform a normal migration the ping doesn't recover anymore, 
but answers repeatedly with: Destination Host Unreachable

Do you know what could be the problem? Where is the difference between a normal 
and live migration and how can the ping after live migrating still work, but 
after a normal migration not?

Thanks a lot!
Regards, Adnan
_______________________________________________
libvirt-users mailing list
libvirt-users@redhat.com
https://www.redhat.com/mailman/listinfo/libvirt-users

Reply via email to