Thanks, Fredy for your great help. Setting the Banner and PrintMotd options
on all 3 nodes helped me to succeed with the installation.
Am Fr., 14. Feb. 2020 um 16:23 Uhr schrieb Fredy Sanchez <
[email protected]>:

> Banner none
> PrintMotd no
>
> # systemctl restart sshd
>

That should be fixed in the ovirt-node images.


> If gluster installed successfully, you don't have to reinstall it.
> Just run the hyperconverged install again from cockpit, and it will detect
> the existing gluster install, and ask you if you want to re-use it;
> re-using worked for me. Only thing I'd point out here is that gluster
> didn't enable in my servers automagically; I had to enable it and start it
> by hand before cockpit picked it up.
> # systemctl enable glusterd --now
> # systemctl status glusterd
>
> Gluster was running fine for me. For me that was not needed.

Also,
> # tail -f /var/log/secure
> while the install is going will help you see if there's a problem with
> ssh, other than the banners.
>
> --
> Fredy
>
> On Fri, Feb 14, 2020 at 9:32 AM Florian Nolden <[email protected]>
> wrote:
>
>>
>> Am Fr., 14. Feb. 2020 um 12:21 Uhr schrieb Fredy Sanchez <
>> [email protected]>:
>>
>>> Hi Florian,
>>>
>>
>>> In my case, Didi's suggestions got me thinking, and I ultimately traced
>>> this to the ssh banners; they must be disabled. You can do this in
>>> sshd_config. I do think that logging could be better for this issue, and
>>> that the host up check should incorporate things other than ssh, even if
>>> just a ping. Good luck.
>>>
>>> Hi Fredy,
>>
>> thanks for the reply.
>>
>> I just have to uncomment "Banners none" in the /etc/ssh/sshd_config on
>> all 3 nodes, and run redeploy in the cockpit?
>> Or have you also reinstalled the nodes and the gluster storage?
>>
>>> --
>>> Fredy
>>>
>>> On Fri, Feb 14, 2020, 4:55 AM Florian Nolden <[email protected]>
>>> wrote:
>>>
>>>> I'also stuck with that issue.
>>>>
>>>> I have
>>>> 3x  HP ProLiant DL360 G7
>>>>
>>>> 1x 1gbit => as control network
>>>> 3x 1gbit => bond0 as Lan
>>>> 2x 10gbit => bond1 as gluster network
>>>>
>>>> I installed on all 3 servers Ovirt Node 4.3.8
>>>> configured the networks using cockpit.
>>>> followed this guide for the gluster setup with cockpit:
>>>> https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hyperconverged.html
>>>>
>>>> the installed the hosted engine with cockpit ->:
>>>>
>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Wait for the host to be up]
>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": {"ovirt_hosts": 
>>>> [{"address": "x-c01-n01.lan.xilloc.com", "affinity_labels": [], 
>>>> "auto_numa_status": "unknown", "certificate": {"organization": 
>>>> "lan.xilloc.com", "subject": 
>>>> "O=lan.xilloc.com,CN=x-c01-n01.lan.xilloc.com"}, "cluster": {"href": 
>>>> "/ovirt-engine/api/clusters/3dff6890-4e7b-11ea-90cb-00163e6a7afe", "id": 
>>>> "3dff6890-4e7b-11ea-90cb-00163e6a7afe"}, "comment": "", "cpu": {"speed": 
>>>> 0.0, "topology": {}}, "device_passthrough": {"enabled": false}, "devices": 
>>>> [], "external_network_provider_configurations": [], "external_status": 
>>>> "ok", "hardware_information": {"supported_rng_sources": []}, "hooks": [], 
>>>> "href": "/ovirt-engine/api/hosts/ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", 
>>>> "id": "ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", "katello_errata": [], 
>>>> "kdump_status": "unknown", "ksm": {"enabled": false}, 
>>>> "max_scheduling_memory": 0, "memory": 0, "name": 
>>>> "x-c01-n01.lan.xilloc.com", "network_attachments": [], "nics": [], 
>>>> "numa_nodes": [], "numa_supported": false, "os": {"custom_kernel_cmdline": 
>>>> ""}, "permissions": [], "port": 54321, "power_management": 
>>>> {"automatic_pm_enabled": true, "enabled": false, "kdump_detection": true, 
>>>> "pm_proxies": []}, "protocol": "stomp", "se_linux": {}, "spm": 
>>>> {"priority": 5, "status": "none"}, "ssh": {"fingerprint": 
>>>> "SHA256:lWc/BuE5WukHd95WwfmFW2ee8VPJ2VugvJeI0puMlh4", "port": 22}, 
>>>> "statistics": [], "status": "non_responsive", 
>>>> "storage_connection_extensions": [], "summary": {"total": 0}, "tags": [], 
>>>> "transparent_huge_pages": {"enabled": false}, "type": "ovirt_node", 
>>>> "unmanaged_networks": [], "update_available": false, "vgpu_placement": 
>>>> "consolidated"}]}, "attempts": 120, "changed": false, "deprecations": 
>>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to 
>>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts", 
>>>> "version": "2.13"}]}
>>>>
>>>>
>>>>
>>>> What is the best approach now to install a Ovirt Hostedengine?
>>>>
>>>>
>>>> Kind regards,
>>>>
>>>> *Florian Nolden*
>>>>
>>>> *Head of IT at Xilloc Medical B.V.*
>>>>
>>>> www.xilloc.com* “Get aHead with patient specific implants” *
>>>>
>>>> Xilloc Medical B.V., Urmonderbaan 22
>>>> <https://maps.google.com/?q=Xilloc%20Medical+B.V.,+Urmonderbaan+22&entry=gmail&source=g>
>>>>  Gate
>>>> 2, Building 110, 6167 RD Sittard-Geleen
>>>>
>>>> —————————————————————————————————————
>>>>
>>>> Disclaimer: The content of this e-mail, including any attachments, are
>>>> confidential and are intended for the sole use of the individual or entity
>>>> to which it is addressed. If you have received it by mistake please let us
>>>> know by reply and then delete it from your system. Any distribution,
>>>> copying or dissemination of this message is expected to conform to all
>>>> legal stipulations governing the use of information.
>>>>
>>>>
>>>> Am Mo., 27. Jan. 2020 um 07:56 Uhr schrieb Yedidyah Bar David <
>>>> [email protected]>:
>>>>
>>>>> On Sun, Jan 26, 2020 at 8:45 PM Fredy Sanchez <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> *Hi all,*
>>>>>>
>>>>>> *[root@bric-ovirt-1 ~]# cat /etc/*release**
>>>>>> CentOS Linux release 7.7.1908 (Core)
>>>>>> *[root@bric-ovirt-1 ~]# yum info ovirt-engine-appliance*
>>>>>> Installed Packages
>>>>>> Name        : ovirt-engine-appliance
>>>>>> Arch        : x86_64
>>>>>> Version     : 4.3
>>>>>> Release     : 20191121.1.el7
>>>>>> Size        : 1.0 G
>>>>>> Repo        : installed
>>>>>> From repo   : ovirt-4.3
>>>>>>
>>>>>> *Same situation as
>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1787267
>>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1787267>. The error message
>>>>>> almost everywhere is some red herring message about ansible*
>>>>>>
>>>>>
>>>>> You are right that it's misleading, but were the errors below the only
>>>>> ones you got from ansible?
>>>>>
>>>>>
>>>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Wait for the host to be
>>>>>> up]
>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts":
>>>>>> {"ovirt_hosts": []}, "attempts": 120, "changed": false, "deprecations":
>>>>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to
>>>>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts",
>>>>>> "version": "2.13"}]}
>>>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Notify the user about a
>>>>>> failure]
>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
>>>>>> "The system may not be provisioned according to the playbook results:
>>>>>> please check the logs for the issue, fix accordingly or re-deploy from
>>>>>> scratch.\n"}
>>>>>> [ ERROR ] Failed to execute stage 'Closing up': Failed executing
>>>>>> ansible-playbook
>>>>>> [ INFO  ] Stage: Termination
>>>>>> [ ERROR ] Hosted Engine deployment failed: please check the logs for
>>>>>> the issue, fix accordingly or re-deploy from scratch.
>>>>>>           Log file is located at
>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20200126170315-req4qb.log
>>>>>>
>>>>>> *But the "real" problem seems to be SSH related, as you can see below*
>>>>>>
>>>>>
>>>>> Indeed
>>>>>
>>>>>
>>>>>> *[root@bric-ovirt-1 ovirt-engine]# pwd*
>>>>>>
>>>>>> /var/log/ovirt-hosted-engine-setup/engine-logs-2020-01-26T17:19:28Z/ovirt-engine
>>>>>> *[root@bric-ovirt-1 ovirt-engine]# grep -i error engine.log*
>>>>>> 2020-01-26 17:26:50,178Z ERROR
>>>>>> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1)
>>>>>> [2341fd23-f0c7-4f1c-ad48-88af20c2d04b] Failed to establish session with
>>>>>> host 'bric-ovirt-1.corp.modmed.com': SSH session closed during
>>>>>> connection '[email protected]'
>>>>>>
>>>>>
>>>>> Please check/share the entire portion of engine.log, from where it
>>>>> starts to try to ssh til it gives up.
>>>>>
>>>>>
>>>>>> 2020-01-26 17:26:50,205Z ERROR
>>>>>> [org.ovirt.engine.api.restapi.resource.AbstractBackendResource] (default
>>>>>> task-1) [] Operation Failed: [Cannot add Host. Connecting to host via SSH
>>>>>> has failed, verify that the host is reachable (IP address, routable 
>>>>>> address
>>>>>> etc.) You may refer to the engine.log file for further details.]
>>>>>>
>>>>>> *The funny thing is that the engine can indeed ssh to bric-ovirt-1
>>>>>> (physical host). See below*
>>>>>>
>>>>>> *[root@bric-ovirt-1 ovirt-hosted-engine-setup]# cat /etc/hosts*
>>>>>> 192.168.1.52 bric-ovirt-engine.corp.modmed.com # temporary entry
>>>>>> added by hosted-engine-setup for the bootstrap VM
>>>>>> 127.0.0.1   localhost localhost.localdomain localhost4
>>>>>> localhost4.localdomain4
>>>>>> #::1         localhost localhost.localdomain localhost6
>>>>>> localhost6.localdomain6
>>>>>> 10.130.0.50 bric-ovirt-engine bric-ovirt-engine.corp.modmed.com
>>>>>> 10.130.0.51 bric-ovirt-1 bric-ovirt-1.corp.modmed.com
>>>>>> 10.130.0.52 bric-ovirt-2 bric-ovirt-2.corp.modmed.com
>>>>>> 10.130.0.53 bric-ovirt-3 bric-ovirt-3.corp.modmed.com
>>>>>> 192.168.0.1 bric-ovirt-1gluster bric-ovirt-1gluster.corp.modmed.com
>>>>>> 192.168.0.2 bric-ovirt-2gluster bric-ovirt-2gluster.corp.modmed.com
>>>>>> 192.168.0.3 bric-ovirt-3gluster bric-ovirt-3gluster.corp.modmed.com
>>>>>> [root@bric-ovirt-1 ovirt-hosted-engine-setup]#
>>>>>>
>>>>>> *[root@bric-ovirt-1 ~]# ssh 192.168.1.52*
>>>>>> Last login: Sun Jan 26 17:55:20 2020 from 192.168.1.1
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1*
>>>>>> Password:
>>>>>> Password:
>>>>>> Last failed login: Sun Jan 26 18:17:16 UTC 2020 from 192.168.1.52 on
>>>>>> ssh:notty
>>>>>> There was 1 failed login attempt since the last successful login.
>>>>>> Last login: Sun Jan 26 18:16:46 2020
>>>>>> ###################################################################
>>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED                #
>>>>>> #                                                                 #
>>>>>> # This system is the property of Modernizing Medicine, Inc.       #
>>>>>> # It is for authorized Company business purposes only.            #
>>>>>> # All connections are monitored and recorded.                     #
>>>>>> # Disconnect IMMEDIATELY if you are not an authorized user!       #
>>>>>> ###################################################################
>>>>>> [root@bric-ovirt-1 ~]#
>>>>>> [root@bric-ovirt-1 ~]#
>>>>>> [root@bric-ovirt-1 ~]# exit
>>>>>> logout
>>>>>> Connection to bric-ovirt-1 closed.
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1.corp.modmed.com
>>>>>> <http://bric-ovirt-1.corp.modmed.com>*
>>>>>> Password:
>>>>>> Last login: Sun Jan 26 18:17:22 2020 from 192.168.1.52
>>>>>> ###################################################################
>>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED                #
>>>>>> #                                                                 #
>>>>>> # This system is the property of Modernizing Medicine, Inc.       #
>>>>>> # It is for authorized Company business purposes only.            #
>>>>>> # All connections are monitored and recorded.                     #
>>>>>> # Disconnect IMMEDIATELY if you are not an authorized user!       #
>>>>>> ###################################################################
>>>>>>
>>>>>
>>>>> Can you please try this, from the engine machine:
>>>>>
>>>>> ssh [email protected] true
>>>>>
>>>>> If this outputs the above "PROHIBITED" note, you'll have to configure
>>>>> your
>>>>> scripts etc. to not output it on non-interactive shells. Otherwise,
>>>>> this
>>>>> confuses the engine - it can't really distinguish between your own
>>>>> output
>>>>> and the output of the commands it runs there.
>>>>>
>>>>>
>>>>>> [root@bric-ovirt-1 ~]# exit
>>>>>> logout
>>>>>> Connection to bric-ovirt-1.corp.modmed.com closed.
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> [root@bric-ovirt-engine ~]#
>>>>>> [root@bric-ovirt-engine ~]# exit
>>>>>> logout
>>>>>> Connection to 192.168.1.52 closed.
>>>>>> [root@bric-ovirt-1 ~]#
>>>>>>
>>>>>> *So, what gives? I already disabled all ssh security in the physical
>>>>>> host, and whitelisted all potential IPs from the engine using firewalld.
>>>>>> Regardless, the engine can ssh to the host as root :-(. Is there maybe
>>>>>> another user that's used for the "Wait for the host to be  up" SSH test?
>>>>>> Yes, I tried both passwords and certificates.*
>>>>>>
>>>>>
>>>>> No, that's root. You can also see that in the log.
>>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>> *Maybe what's really happening is that engine is not getting the
>>>>>> right IP? bric-ovirt-engine is supposed to get 10.130.0.50, instead it
>>>>>> never gets there, getting 192.168.1.52 from virbr0 in bric-ovirt-1. See
>>>>>> below.*
>>>>>>
>>>>>
>>>>> That's by design. For details, if interested, see "Hosted Engine 4.3
>>>>> Deep Dive" presentation:
>>>>>
>>>>> https://www.ovirt.org/community/get-involved/resources/slide-decks.html
>>>>>
>>>>>
>>>>>>
>>>>>>  --== HOST NETWORK CONFIGURATION ==--
>>>>>>           Please indicate the gateway IP address [10.130.0.1]
>>>>>>           Please indicate a nic to set ovirtmgmt bridge on: (p4p1,
>>>>>> p5p1) [p4p1]:
>>>>>> --== VM CONFIGURATION ==--
>>>>>> You may specify a unicast MAC address for the VM or accept a randomly
>>>>>> generated default [00:16:3e:17:1d:f8]:
>>>>>>           How should the engine VM network be configured (DHCP,
>>>>>> Static)[DHCP]? static
>>>>>>           Please enter the IP address to be used for the engine VM
>>>>>> []: 10.130.0.50
>>>>>> [ INFO  ] The engine VM will be configured to use 10.130.0.50/25
>>>>>>           Please provide a comma-separated list (max 3) of IP
>>>>>> addresses of domain name servers for the engine VM
>>>>>>           Engine VM DNS (leave it empty to skip)
>>>>>> [10.130.0.2,10.130.0.3]:
>>>>>>           Add lines for the appliance itself and for this host to
>>>>>> /etc/hosts on the engine VM?
>>>>>>           Note: ensuring that this host could resolve the engine VM
>>>>>> hostname is still up to you
>>>>>>           (Yes, No)[No] Yes
>>>>>>
>>>>>> *[root@bric-ovirt-1 ~]# ip addr*
>>>>>> 3: p4p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>>>>>> group default qlen 1000
>>>>>>     link/ether 00:0a:f7:f1:c6:80 brd ff:ff:ff:ff:ff:ff
>>>>>>     inet 10.130.0.51/25 brd 10.130.0.127 scope global noprefixroute
>>>>>> p4p1
>>>>>>        valid_lft forever preferred_lft forever
>>>>>> 28: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>>>>>> state UP group default qlen 1000
>>>>>>     link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff
>>>>>>     inet 192.168.1.1/24 brd 192.168.1.255 scope global virbr0
>>>>>>        valid_lft forever preferred_lft forever
>>>>>> 29: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast
>>>>>> master virbr0 state DOWN group default qlen 1000
>>>>>>     link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff
>>>>>> 30: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>>> pfifo_fast master virbr0 state UNKNOWN group default qlen 1000
>>>>>>     link/ether fe:16:3e:17:1d:f8 brd ff:ff:ff:ff:ff:ff
>>>>>>
>>>>>> *The newly created engine VM does remain up even after hosted-engine
>>>>>> --deploy errors out; just at the wrong IP. I haven't been able to make it
>>>>>> get its real IP.*
>>>>>>
>>>>>
>>>>> This happens only after the real engine VM is created, connected to
>>>>> the correct network.
>>>>>
>>>>> The current engine vm you see is a libvirt VM connected to its default
>>>>> (internal) network.
>>>>>
>>>>>
>>>>>> * At any rate, thank you very much for taking a look at my very long
>>>>>> email. Any and all help would be really appreciated.*
>>>>>>
>>>>>
>>>>> Good luck and best regards,
>>>>> --
>>>>> Didi
>>>>> _______________________________________________
>>>>> Users mailing list -- [email protected]
>>>>> To unsubscribe send an email to [email protected]
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/AZFPSDPBK3BJUB2NESCOWQ7FQT572Y5I/
>>>>>
>>>>
>>> *CONFIDENTIALITY NOTICE:* This e-mail message may contain material
>>> protected by the Health Insurance Portability and Accountability Act of
>>> 1996 and its implementing regulations and other state and federal laws and
>>> legal privileges. This message is only for the personal and confidential
>>> use of the individuals or organization to whom the message is addressed. If
>>> you are an unintended recipient, you have received this message in error,
>>> and any reading, distributing, copying or disclosure is unauthorized and
>>> strictly prohibited.  All recipients are hereby notified that any
>>> unauthorized receipt does not waive any confidentiality obligations or
>>> privileges. If you have received this message in error, please notify the
>>> sender immediately at the above email address and confirm that you have
>>> deleted or destroyed the message.
>>
>>
>
>
>
> *CONFIDENTIALITY NOTICE:* This e-mail message may contain material
> protected by the Health Insurance Portability and Accountability Act of
> 1996 and its implementing regulations and other state and federal laws and
> legal privileges. This message is only for the personal and confidential
> use of the individuals or organization to whom the message is addressed. If
> you are an unintended recipient, you have received this message in error,
> and any reading, distributing, copying or disclosure is unauthorized and
> strictly prohibited.  All recipients are hereby notified that any
> unauthorized receipt does not waive any confidentiality obligations or
> privileges. If you have received this message in error, please notify the
> sender immediately at the above email address and confirm that you have
> deleted or destroyed the message.
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/CIPHZSL3KO4X53O45BRPIKOIQPHJA6LF/

Reply via email to