Hi,

I think this is a FAQ, but I seem to be blind enough to miss any clue
about this issue.

I have recently (test-)migrated our KVM-based virtualisation
infrastructure from one-standard-bridge-per-VLAN to a fully VLAN-aware
openvswitch (Debian testing, which means kernel 3.2.32 with
openvswitch-datapath-dkms 1.4.2) because it feels much more natural to
use.

I don't think the actual configuration matters a lot, but here it comes

- two GigE ports in active-standby bond0 to upstream switches
- two GigE ports in LACP bond1 to partner

I have ruled out any CPU load issues mentioned in the FAQ like loops and
broadcast storms.

My problem is that when I move some high traffic VMs to this host the
CPU usage of ovs-vswitchd is massively increased. I'm reasonably sure
this is due to the fact that there seems to be a flow setup for each L4
connection through that vswitch, which in this case is highly
unfortunate (firewall VM with a lot of concurrent sessions).

root@virt1:~# ovs-dpctl show       
system@br0:
        lookups: hit:5200011861 missed:829071115 lost:279074
        flows: 34392

I have already set other_config:flow-eviction-threshold=100000,
otherwise it would be a lot worse. But I still have a CPU load that is
several times higher than it was before with the standard bridge, and I
still see occasional loss during high traffic periods.

I don't actually need all that fancy OpenFlow stuff, so in my case a
traditional switching based on MAC learning and the destination MAC
would suffice. I cannot find any information whether that is possible.

Thanks,
Bernhard


_______________________________________________
discuss mailing list
discuss@openvswitch.org
http://openvswitch.org/mailman/listinfo/discuss

Reply via email to