Thanks, Fredy for your great help. Setting the Banner and PrintMotd options on all 3 nodes helped me to succeed with the installation. Am Fr., 14. Feb. 2020 um 16:23 Uhr schrieb Fredy Sanchez < [email protected]>:
> Banner none > PrintMotd no > > # systemctl restart sshd > That should be fixed in the ovirt-node images. > If gluster installed successfully, you don't have to reinstall it. > Just run the hyperconverged install again from cockpit, and it will detect > the existing gluster install, and ask you if you want to re-use it; > re-using worked for me. Only thing I'd point out here is that gluster > didn't enable in my servers automagically; I had to enable it and start it > by hand before cockpit picked it up. > # systemctl enable glusterd --now > # systemctl status glusterd > > Gluster was running fine for me. For me that was not needed. Also, > # tail -f /var/log/secure > while the install is going will help you see if there's a problem with > ssh, other than the banners. > > -- > Fredy > > On Fri, Feb 14, 2020 at 9:32 AM Florian Nolden <[email protected]> > wrote: > >> >> Am Fr., 14. Feb. 2020 um 12:21 Uhr schrieb Fredy Sanchez < >> [email protected]>: >> >>> Hi Florian, >>> >> >>> In my case, Didi's suggestions got me thinking, and I ultimately traced >>> this to the ssh banners; they must be disabled. You can do this in >>> sshd_config. I do think that logging could be better for this issue, and >>> that the host up check should incorporate things other than ssh, even if >>> just a ping. Good luck. >>> >>> Hi Fredy, >> >> thanks for the reply. >> >> I just have to uncomment "Banners none" in the /etc/ssh/sshd_config on >> all 3 nodes, and run redeploy in the cockpit? >> Or have you also reinstalled the nodes and the gluster storage? >> >>> -- >>> Fredy >>> >>> On Fri, Feb 14, 2020, 4:55 AM Florian Nolden <[email protected]> >>> wrote: >>> >>>> I'also stuck with that issue. >>>> >>>> I have >>>> 3x HP ProLiant DL360 G7 >>>> >>>> 1x 1gbit => as control network >>>> 3x 1gbit => bond0 as Lan >>>> 2x 10gbit => bond1 as gluster network >>>> >>>> I installed on all 3 servers Ovirt Node 4.3.8 >>>> configured the networks using cockpit. >>>> followed this guide for the gluster setup with cockpit: >>>> https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hyperconverged.html >>>> >>>> the installed the hosted engine with cockpit ->: >>>> >>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Wait for the host to be up] >>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": {"ovirt_hosts": >>>> [{"address": "x-c01-n01.lan.xilloc.com", "affinity_labels": [], >>>> "auto_numa_status": "unknown", "certificate": {"organization": >>>> "lan.xilloc.com", "subject": >>>> "O=lan.xilloc.com,CN=x-c01-n01.lan.xilloc.com"}, "cluster": {"href": >>>> "/ovirt-engine/api/clusters/3dff6890-4e7b-11ea-90cb-00163e6a7afe", "id": >>>> "3dff6890-4e7b-11ea-90cb-00163e6a7afe"}, "comment": "", "cpu": {"speed": >>>> 0.0, "topology": {}}, "device_passthrough": {"enabled": false}, "devices": >>>> [], "external_network_provider_configurations": [], "external_status": >>>> "ok", "hardware_information": {"supported_rng_sources": []}, "hooks": [], >>>> "href": "/ovirt-engine/api/hosts/ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", >>>> "id": "ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", "katello_errata": [], >>>> "kdump_status": "unknown", "ksm": {"enabled": false}, >>>> "max_scheduling_memory": 0, "memory": 0, "name": >>>> "x-c01-n01.lan.xilloc.com", "network_attachments": [], "nics": [], >>>> "numa_nodes": [], "numa_supported": false, "os": {"custom_kernel_cmdline": >>>> ""}, "permissions": [], "port": 54321, "power_management": >>>> {"automatic_pm_enabled": true, "enabled": false, "kdump_detection": true, >>>> "pm_proxies": []}, "protocol": "stomp", "se_linux": {}, "spm": >>>> {"priority": 5, "status": "none"}, "ssh": {"fingerprint": >>>> "SHA256:lWc/BuE5WukHd95WwfmFW2ee8VPJ2VugvJeI0puMlh4", "port": 22}, >>>> "statistics": [], "status": "non_responsive", >>>> "storage_connection_extensions": [], "summary": {"total": 0}, "tags": [], >>>> "transparent_huge_pages": {"enabled": false}, "type": "ovirt_node", >>>> "unmanaged_networks": [], "update_available": false, "vgpu_placement": >>>> "consolidated"}]}, "attempts": 120, "changed": false, "deprecations": >>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to >>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts", >>>> "version": "2.13"}]} >>>> >>>> >>>> >>>> What is the best approach now to install a Ovirt Hostedengine? >>>> >>>> >>>> Kind regards, >>>> >>>> *Florian Nolden* >>>> >>>> *Head of IT at Xilloc Medical B.V.* >>>> >>>> www.xilloc.com* “Get aHead with patient specific implants” * >>>> >>>> Xilloc Medical B.V., Urmonderbaan 22 >>>> <https://maps.google.com/?q=Xilloc%20Medical+B.V.,+Urmonderbaan+22&entry=gmail&source=g> >>>> Gate >>>> 2, Building 110, 6167 RD Sittard-Geleen >>>> >>>> ————————————————————————————————————— >>>> >>>> Disclaimer: The content of this e-mail, including any attachments, are >>>> confidential and are intended for the sole use of the individual or entity >>>> to which it is addressed. If you have received it by mistake please let us >>>> know by reply and then delete it from your system. Any distribution, >>>> copying or dissemination of this message is expected to conform to all >>>> legal stipulations governing the use of information. >>>> >>>> >>>> Am Mo., 27. Jan. 2020 um 07:56 Uhr schrieb Yedidyah Bar David < >>>> [email protected]>: >>>> >>>>> On Sun, Jan 26, 2020 at 8:45 PM Fredy Sanchez < >>>>> [email protected]> wrote: >>>>> >>>>>> *Hi all,* >>>>>> >>>>>> *[root@bric-ovirt-1 ~]# cat /etc/*release** >>>>>> CentOS Linux release 7.7.1908 (Core) >>>>>> *[root@bric-ovirt-1 ~]# yum info ovirt-engine-appliance* >>>>>> Installed Packages >>>>>> Name : ovirt-engine-appliance >>>>>> Arch : x86_64 >>>>>> Version : 4.3 >>>>>> Release : 20191121.1.el7 >>>>>> Size : 1.0 G >>>>>> Repo : installed >>>>>> From repo : ovirt-4.3 >>>>>> >>>>>> *Same situation as >>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1787267 >>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1787267>. The error message >>>>>> almost everywhere is some red herring message about ansible* >>>>>> >>>>> >>>>> You are right that it's misleading, but were the errors below the only >>>>> ones you got from ansible? >>>>> >>>>> >>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Wait for the host to be >>>>>> up] >>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": >>>>>> {"ovirt_hosts": []}, "attempts": 120, "changed": false, "deprecations": >>>>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to >>>>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts", >>>>>> "version": "2.13"}]} >>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Notify the user about a >>>>>> failure] >>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": >>>>>> "The system may not be provisioned according to the playbook results: >>>>>> please check the logs for the issue, fix accordingly or re-deploy from >>>>>> scratch.\n"} >>>>>> [ ERROR ] Failed to execute stage 'Closing up': Failed executing >>>>>> ansible-playbook >>>>>> [ INFO ] Stage: Termination >>>>>> [ ERROR ] Hosted Engine deployment failed: please check the logs for >>>>>> the issue, fix accordingly or re-deploy from scratch. >>>>>> Log file is located at >>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20200126170315-req4qb.log >>>>>> >>>>>> *But the "real" problem seems to be SSH related, as you can see below* >>>>>> >>>>> >>>>> Indeed >>>>> >>>>> >>>>>> *[root@bric-ovirt-1 ovirt-engine]# pwd* >>>>>> >>>>>> /var/log/ovirt-hosted-engine-setup/engine-logs-2020-01-26T17:19:28Z/ovirt-engine >>>>>> *[root@bric-ovirt-1 ovirt-engine]# grep -i error engine.log* >>>>>> 2020-01-26 17:26:50,178Z ERROR >>>>>> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) >>>>>> [2341fd23-f0c7-4f1c-ad48-88af20c2d04b] Failed to establish session with >>>>>> host 'bric-ovirt-1.corp.modmed.com': SSH session closed during >>>>>> connection '[email protected]' >>>>>> >>>>> >>>>> Please check/share the entire portion of engine.log, from where it >>>>> starts to try to ssh til it gives up. >>>>> >>>>> >>>>>> 2020-01-26 17:26:50,205Z ERROR >>>>>> [org.ovirt.engine.api.restapi.resource.AbstractBackendResource] (default >>>>>> task-1) [] Operation Failed: [Cannot add Host. Connecting to host via SSH >>>>>> has failed, verify that the host is reachable (IP address, routable >>>>>> address >>>>>> etc.) You may refer to the engine.log file for further details.] >>>>>> >>>>>> *The funny thing is that the engine can indeed ssh to bric-ovirt-1 >>>>>> (physical host). See below* >>>>>> >>>>>> *[root@bric-ovirt-1 ovirt-hosted-engine-setup]# cat /etc/hosts* >>>>>> 192.168.1.52 bric-ovirt-engine.corp.modmed.com # temporary entry >>>>>> added by hosted-engine-setup for the bootstrap VM >>>>>> 127.0.0.1 localhost localhost.localdomain localhost4 >>>>>> localhost4.localdomain4 >>>>>> #::1 localhost localhost.localdomain localhost6 >>>>>> localhost6.localdomain6 >>>>>> 10.130.0.50 bric-ovirt-engine bric-ovirt-engine.corp.modmed.com >>>>>> 10.130.0.51 bric-ovirt-1 bric-ovirt-1.corp.modmed.com >>>>>> 10.130.0.52 bric-ovirt-2 bric-ovirt-2.corp.modmed.com >>>>>> 10.130.0.53 bric-ovirt-3 bric-ovirt-3.corp.modmed.com >>>>>> 192.168.0.1 bric-ovirt-1gluster bric-ovirt-1gluster.corp.modmed.com >>>>>> 192.168.0.2 bric-ovirt-2gluster bric-ovirt-2gluster.corp.modmed.com >>>>>> 192.168.0.3 bric-ovirt-3gluster bric-ovirt-3gluster.corp.modmed.com >>>>>> [root@bric-ovirt-1 ovirt-hosted-engine-setup]# >>>>>> >>>>>> *[root@bric-ovirt-1 ~]# ssh 192.168.1.52* >>>>>> Last login: Sun Jan 26 17:55:20 2020 from 192.168.1.1 >>>>>> [root@bric-ovirt-engine ~]# >>>>>> [root@bric-ovirt-engine ~]# >>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1* >>>>>> Password: >>>>>> Password: >>>>>> Last failed login: Sun Jan 26 18:17:16 UTC 2020 from 192.168.1.52 on >>>>>> ssh:notty >>>>>> There was 1 failed login attempt since the last successful login. >>>>>> Last login: Sun Jan 26 18:16:46 2020 >>>>>> ################################################################### >>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED # >>>>>> # # >>>>>> # This system is the property of Modernizing Medicine, Inc. # >>>>>> # It is for authorized Company business purposes only. # >>>>>> # All connections are monitored and recorded. # >>>>>> # Disconnect IMMEDIATELY if you are not an authorized user! # >>>>>> ################################################################### >>>>>> [root@bric-ovirt-1 ~]# >>>>>> [root@bric-ovirt-1 ~]# >>>>>> [root@bric-ovirt-1 ~]# exit >>>>>> logout >>>>>> Connection to bric-ovirt-1 closed. >>>>>> [root@bric-ovirt-engine ~]# >>>>>> [root@bric-ovirt-engine ~]# >>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1.corp.modmed.com >>>>>> <http://bric-ovirt-1.corp.modmed.com>* >>>>>> Password: >>>>>> Last login: Sun Jan 26 18:17:22 2020 from 192.168.1.52 >>>>>> ################################################################### >>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED # >>>>>> # # >>>>>> # This system is the property of Modernizing Medicine, Inc. # >>>>>> # It is for authorized Company business purposes only. # >>>>>> # All connections are monitored and recorded. # >>>>>> # Disconnect IMMEDIATELY if you are not an authorized user! # >>>>>> ################################################################### >>>>>> >>>>> >>>>> Can you please try this, from the engine machine: >>>>> >>>>> ssh [email protected] true >>>>> >>>>> If this outputs the above "PROHIBITED" note, you'll have to configure >>>>> your >>>>> scripts etc. to not output it on non-interactive shells. Otherwise, >>>>> this >>>>> confuses the engine - it can't really distinguish between your own >>>>> output >>>>> and the output of the commands it runs there. >>>>> >>>>> >>>>>> [root@bric-ovirt-1 ~]# exit >>>>>> logout >>>>>> Connection to bric-ovirt-1.corp.modmed.com closed. >>>>>> [root@bric-ovirt-engine ~]# >>>>>> [root@bric-ovirt-engine ~]# >>>>>> [root@bric-ovirt-engine ~]# exit >>>>>> logout >>>>>> Connection to 192.168.1.52 closed. >>>>>> [root@bric-ovirt-1 ~]# >>>>>> >>>>>> *So, what gives? I already disabled all ssh security in the physical >>>>>> host, and whitelisted all potential IPs from the engine using firewalld. >>>>>> Regardless, the engine can ssh to the host as root :-(. Is there maybe >>>>>> another user that's used for the "Wait for the host to be up" SSH test? >>>>>> Yes, I tried both passwords and certificates.* >>>>>> >>>>> >>>>> No, that's root. You can also see that in the log. >>>>> >>>>> >>>>>> >>>>>> >>>>>> *Maybe what's really happening is that engine is not getting the >>>>>> right IP? bric-ovirt-engine is supposed to get 10.130.0.50, instead it >>>>>> never gets there, getting 192.168.1.52 from virbr0 in bric-ovirt-1. See >>>>>> below.* >>>>>> >>>>> >>>>> That's by design. For details, if interested, see "Hosted Engine 4.3 >>>>> Deep Dive" presentation: >>>>> >>>>> https://www.ovirt.org/community/get-involved/resources/slide-decks.html >>>>> >>>>> >>>>>> >>>>>> --== HOST NETWORK CONFIGURATION ==-- >>>>>> Please indicate the gateway IP address [10.130.0.1] >>>>>> Please indicate a nic to set ovirtmgmt bridge on: (p4p1, >>>>>> p5p1) [p4p1]: >>>>>> --== VM CONFIGURATION ==-- >>>>>> You may specify a unicast MAC address for the VM or accept a randomly >>>>>> generated default [00:16:3e:17:1d:f8]: >>>>>> How should the engine VM network be configured (DHCP, >>>>>> Static)[DHCP]? static >>>>>> Please enter the IP address to be used for the engine VM >>>>>> []: 10.130.0.50 >>>>>> [ INFO ] The engine VM will be configured to use 10.130.0.50/25 >>>>>> Please provide a comma-separated list (max 3) of IP >>>>>> addresses of domain name servers for the engine VM >>>>>> Engine VM DNS (leave it empty to skip) >>>>>> [10.130.0.2,10.130.0.3]: >>>>>> Add lines for the appliance itself and for this host to >>>>>> /etc/hosts on the engine VM? >>>>>> Note: ensuring that this host could resolve the engine VM >>>>>> hostname is still up to you >>>>>> (Yes, No)[No] Yes >>>>>> >>>>>> *[root@bric-ovirt-1 ~]# ip addr* >>>>>> 3: p4p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP >>>>>> group default qlen 1000 >>>>>> link/ether 00:0a:f7:f1:c6:80 brd ff:ff:ff:ff:ff:ff >>>>>> inet 10.130.0.51/25 brd 10.130.0.127 scope global noprefixroute >>>>>> p4p1 >>>>>> valid_lft forever preferred_lft forever >>>>>> 28: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue >>>>>> state UP group default qlen 1000 >>>>>> link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff >>>>>> inet 192.168.1.1/24 brd 192.168.1.255 scope global virbr0 >>>>>> valid_lft forever preferred_lft forever >>>>>> 29: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast >>>>>> master virbr0 state DOWN group default qlen 1000 >>>>>> link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff >>>>>> 30: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>>>>> pfifo_fast master virbr0 state UNKNOWN group default qlen 1000 >>>>>> link/ether fe:16:3e:17:1d:f8 brd ff:ff:ff:ff:ff:ff >>>>>> >>>>>> *The newly created engine VM does remain up even after hosted-engine >>>>>> --deploy errors out; just at the wrong IP. I haven't been able to make it >>>>>> get its real IP.* >>>>>> >>>>> >>>>> This happens only after the real engine VM is created, connected to >>>>> the correct network. >>>>> >>>>> The current engine vm you see is a libvirt VM connected to its default >>>>> (internal) network. >>>>> >>>>> >>>>>> * At any rate, thank you very much for taking a look at my very long >>>>>> email. Any and all help would be really appreciated.* >>>>>> >>>>> >>>>> Good luck and best regards, >>>>> -- >>>>> Didi >>>>> _______________________________________________ >>>>> Users mailing list -- [email protected] >>>>> To unsubscribe send an email to [email protected] >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/[email protected]/message/AZFPSDPBK3BJUB2NESCOWQ7FQT572Y5I/ >>>>> >>>> >>> *CONFIDENTIALITY NOTICE:* This e-mail message may contain material >>> protected by the Health Insurance Portability and Accountability Act of >>> 1996 and its implementing regulations and other state and federal laws and >>> legal privileges. This message is only for the personal and confidential >>> use of the individuals or organization to whom the message is addressed. If >>> you are an unintended recipient, you have received this message in error, >>> and any reading, distributing, copying or disclosure is unauthorized and >>> strictly prohibited. All recipients are hereby notified that any >>> unauthorized receipt does not waive any confidentiality obligations or >>> privileges. If you have received this message in error, please notify the >>> sender immediately at the above email address and confirm that you have >>> deleted or destroyed the message. >> >> > > > > *CONFIDENTIALITY NOTICE:* This e-mail message may contain material > protected by the Health Insurance Portability and Accountability Act of > 1996 and its implementing regulations and other state and federal laws and > legal privileges. This message is only for the personal and confidential > use of the individuals or organization to whom the message is addressed. If > you are an unintended recipient, you have received this message in error, > and any reading, distributing, copying or disclosure is unauthorized and > strictly prohibited. All recipients are hereby notified that any > unauthorized receipt does not waive any confidentiality obligations or > privileges. If you have received this message in error, please notify the > sender immediately at the above email address and confirm that you have > deleted or destroyed the message.
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/CIPHZSL3KO4X53O45BRPIKOIQPHJA6LF/

