Hi Flint, Happy to help.
Right now the list of controller endpoints is pushed at boot time and loaded into the amphora via config drive/nova. In the future we plan to be able to update this list via the amphora API, but it has not been developed yet. I am pretty sure centos is getting the config file as our gate job that runs with centos 7 amphora has been passing. It should be in the same /etc/octavia/amphora-agent.conf location as the ubuntu based amphora. Michael On Tue, Jul 31, 2018 at 10:05 AM Flint WALRUS <gael.ther...@gmail.com> wrote: > > Hi Michael, thanks a lot for that explanation, it’s actually how I envisioned > the flow. > > I’ll have to produce a diagram for my peers understanding, I maybe can share > it with you. > > There is still one point that seems to be a little bit odd to me. > > How the amphora agent know where to find out the healthManagers and worker > services? Is that because the worker is sending the agent some catalog > informations or because we set that at diskimage-create time? > > If so, I think the Centos based amphora is missing the agent.conf because > currently my vms doesn’t have any. > > Once again thanks for your help! > Le mar. 31 juil. 2018 à 18:15, Michael Johnson <johnso...@gmail.com> a écrit : >> >> Hi Flint, >> >> We don't have a logical network diagram at this time (it's still on >> the to-do list), but I can talk you through it. >> >> The Octavia worker, health manager, and housekeeping need to be able >> to reach the amphora (service VM at this point) over the lb-mgmt-net >> on TCP 9443. It knows the amphora IP addresses on the lb-mgmt-net via >> the database and the information we save from the compute driver (I.e. >> what IP was assigned to the instance). >> >> The Octavia API process does not need to be connected to the >> lb-mgmt-net at this time. It only connects the the messaging bus and >> the Octavia database. Provider drivers may have other connectivity >> requirements for the Octavia API. >> >> The amphorae also send UDP packets back to the health manager on port >> 5555. This is the heartbeat packet from the amphora. It contains the >> health and statistics from that amphora. It know it's list of health >> manager endpoints from the configuration file >> "controller_ip_port_list" >> (https://docs.openstack.org/octavia/latest/configuration/configref.html#health_manager.controller_ip_port_list). >> Each amphora will rotate through that list of endpoints to reduce the >> chance of a network split impacting the heartbeat messages. >> >> This is the only traffic that passed over this network. All of it is >> IP based and can be routed (it does not require L2 connectivity). >> >> Michael >> >> On Tue, Jul 31, 2018 at 2:00 AM Flint WALRUS <gael.ther...@gmail.com> wrote: >> > >> > Hi Folks, >> > >> > I'm currently deploying the Octavia component into our testing environment >> > which is based on KOLLA. >> > >> > So far I'm quite enjoying it as it is pretty much straight forward (Except >> > for some documentation pitfalls), but I'm now facing a weird and hard to >> > debug situation. >> > >> > I actually have a hard time to understand how Amphora are communicating >> > back and forth with the Control Plan components. >> > >> > From my understanding, as soon as I create a new LB, the Control Plan is >> > spawning an instance using the configured Octavia Flavor and Image type, >> > attach it to the LB-MGMT-NET and to the user provided subnet. >> > >> > What I think I'm misunderstanding is the discussion that follows between >> > the amphora and the different components such as the >> > HealthManager/HouseKeeper, the API and the Worker. >> > >> > How is the amphora agent able to found my control plan? Is the >> > HealthManager or the Octavia Worker initiating the communication to the >> > Amphora on port 9443 and so give the agent the API/Control plan >> > internalURL? >> > >> > If anyone have a diagram of the workflow I would be more than happy ^^ >> > >> > Thanks a lot in advance to anyone willing to help :D >> > >> > _______________________________________________ >> > OpenStack-operators mailing list >> > OpenStack-operators@lists.openstack.org >> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators _______________________________________________ OpenStack-operators mailing list OpenStack-operators@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators