Hi all,

On Fri, Feb 14, 2020 at 6:45 PM Florian Nolden <[email protected]> wrote:

> Thanks, Fredy for your great help. Setting the Banner and PrintMotd
> options on all 3 nodes helped me to succeed with the installation.
>

Thanks a lot for the report!


> Am Fr., 14. Feb. 2020 um 16:23 Uhr schrieb Fredy Sanchez <
> [email protected]>:
>
>> Banner none
>> PrintMotd no
>>
>> # systemctl restart sshd
>>
>
> That should be fixed in the ovirt-node images.
>

I think I agree. Would you like to open a bug about this?

I wonder what we can/should do with EL7 hosts (non-ovirt-node).

Also need to check how 4.4 behaves - there, host-deploy was fully rewritten
using ansible. No idea how sensitive ansible is to these banners (compared
with otopi, which is very). Adding Dana.

Best regards,


>
>
>> If gluster installed successfully, you don't have to reinstall it.
>> Just run the hyperconverged install again from cockpit, and it will
>> detect the existing gluster install, and ask you if you want to re-use it;
>> re-using worked for me. Only thing I'd point out here is that gluster
>> didn't enable in my servers automagically; I had to enable it and start it
>> by hand before cockpit picked it up.
>> # systemctl enable glusterd --now
>> # systemctl status glusterd
>>
>> Gluster was running fine for me. For me that was not needed.
>
> Also,
>> # tail -f /var/log/secure
>> while the install is going will help you see if there's a problem with
>> ssh, other than the banners.
>>
>> --
>> Fredy
>>
>> On Fri, Feb 14, 2020 at 9:32 AM Florian Nolden <[email protected]>
>> wrote:
>>
>>>
>>> Am Fr., 14. Feb. 2020 um 12:21 Uhr schrieb Fredy Sanchez <
>>> [email protected]>:
>>>
>>>> Hi Florian,
>>>>
>>>
>>>> In my case, Didi's suggestions got me thinking, and I ultimately traced
>>>> this to the ssh banners; they must be disabled. You can do this in
>>>> sshd_config. I do think that logging could be better for this issue, and
>>>> that the host up check should incorporate things other than ssh, even if
>>>> just a ping. Good luck.
>>>>
>>>> Hi Fredy,
>>>
>>> thanks for the reply.
>>>
>>> I just have to uncomment "Banners none" in the /etc/ssh/sshd_config on
>>> all 3 nodes, and run redeploy in the cockpit?
>>> Or have you also reinstalled the nodes and the gluster storage?
>>>
>>>> --
>>>> Fredy
>>>>
>>>> On Fri, Feb 14, 2020, 4:55 AM Florian Nolden <[email protected]>
>>>> wrote:
>>>>
>>>>> I'also stuck with that issue.
>>>>>
>>>>> I have
>>>>> 3x  HP ProLiant DL360 G7
>>>>>
>>>>> 1x 1gbit => as control network
>>>>> 3x 1gbit => bond0 as Lan
>>>>> 2x 10gbit => bond1 as gluster network
>>>>>
>>>>> I installed on all 3 servers Ovirt Node 4.3.8
>>>>> configured the networks using cockpit.
>>>>> followed this guide for the gluster setup with cockpit:
>>>>> https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hyperconverged.html
>>>>>
>>>>> the installed the hosted engine with cockpit ->:
>>>>>
>>>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Wait for the host to be up]
>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": 
>>>>> {"ovirt_hosts": [{"address": "x-c01-n01.lan.xilloc.com", 
>>>>> "affinity_labels": [], "auto_numa_status": "unknown", "certificate": 
>>>>> {"organization": "lan.xilloc.com", "subject": 
>>>>> "O=lan.xilloc.com,CN=x-c01-n01.lan.xilloc.com"}, "cluster": {"href": 
>>>>> "/ovirt-engine/api/clusters/3dff6890-4e7b-11ea-90cb-00163e6a7afe", "id": 
>>>>> "3dff6890-4e7b-11ea-90cb-00163e6a7afe"}, "comment": "", "cpu": {"speed": 
>>>>> 0.0, "topology": {}}, "device_passthrough": {"enabled": false}, 
>>>>> "devices": [], "external_network_provider_configurations": [], 
>>>>> "external_status": "ok", "hardware_information": 
>>>>> {"supported_rng_sources": []}, "hooks": [], "href": 
>>>>> "/ovirt-engine/api/hosts/ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", "id": 
>>>>> "ded7aa60-4a5e-456e-b899-dd7fc25cc7b3", "katello_errata": [], 
>>>>> "kdump_status": "unknown", "ksm": {"enabled": false}, 
>>>>> "max_scheduling_memory": 0, "memory": 0, "name": 
>>>>> "x-c01-n01.lan.xilloc.com", "network_attachments": [], "nics": [], 
>>>>> "numa_nodes": [], "numa_supported": false, "os": 
>>>>> {"custom_kernel_cmdline": ""}, "permissions": [], "port": 54321, 
>>>>> "power_management": {"automatic_pm_enabled": true, "enabled": false, 
>>>>> "kdump_detection": true, "pm_proxies": []}, "protocol": "stomp", 
>>>>> "se_linux": {}, "spm": {"priority": 5, "status": "none"}, "ssh": 
>>>>> {"fingerprint": "SHA256:lWc/BuE5WukHd95WwfmFW2ee8VPJ2VugvJeI0puMlh4", 
>>>>> "port": 22}, "statistics": [], "status": "non_responsive", 
>>>>> "storage_connection_extensions": [], "summary": {"total": 0}, "tags": [], 
>>>>> "transparent_huge_pages": {"enabled": false}, "type": "ovirt_node", 
>>>>> "unmanaged_networks": [], "update_available": false, "vgpu_placement": 
>>>>> "consolidated"}]}, "attempts": 120, "changed": false, "deprecations": 
>>>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to 
>>>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts", 
>>>>> "version": "2.13"}]}
>>>>>
>>>>>
>>>>>
>>>>> What is the best approach now to install a Ovirt Hostedengine?
>>>>>
>>>>>
>>>>> Kind regards,
>>>>>
>>>>> *Florian Nolden*
>>>>>
>>>>> *Head of IT at Xilloc Medical B.V.*
>>>>>
>>>>> www.xilloc.com* “Get aHead with patient specific implants” *
>>>>>
>>>>> Xilloc Medical B.V., Urmonderbaan 22
>>>>> <https://maps.google.com/?q=Xilloc%20Medical+B.V.,+Urmonderbaan+22&entry=gmail&source=g>
>>>>>  Gate
>>>>> 2, Building 110, 6167 RD Sittard-Geleen
>>>>>
>>>>> —————————————————————————————————————
>>>>>
>>>>> Disclaimer: The content of this e-mail, including any attachments, are
>>>>> confidential and are intended for the sole use of the individual or entity
>>>>> to which it is addressed. If you have received it by mistake please let us
>>>>> know by reply and then delete it from your system. Any distribution,
>>>>> copying or dissemination of this message is expected to conform to all
>>>>> legal stipulations governing the use of information.
>>>>>
>>>>>
>>>>> Am Mo., 27. Jan. 2020 um 07:56 Uhr schrieb Yedidyah Bar David <
>>>>> [email protected]>:
>>>>>
>>>>>> On Sun, Jan 26, 2020 at 8:45 PM Fredy Sanchez <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> *Hi all,*
>>>>>>>
>>>>>>> *[root@bric-ovirt-1 ~]# cat /etc/*release**
>>>>>>> CentOS Linux release 7.7.1908 (Core)
>>>>>>> *[root@bric-ovirt-1 ~]# yum info ovirt-engine-appliance*
>>>>>>> Installed Packages
>>>>>>> Name        : ovirt-engine-appliance
>>>>>>> Arch        : x86_64
>>>>>>> Version     : 4.3
>>>>>>> Release     : 20191121.1.el7
>>>>>>> Size        : 1.0 G
>>>>>>> Repo        : installed
>>>>>>> From repo   : ovirt-4.3
>>>>>>>
>>>>>>> *Same situation as
>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1787267
>>>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1787267>. The error message
>>>>>>> almost everywhere is some red herring message about ansible*
>>>>>>>
>>>>>>
>>>>>> You are right that it's misleading, but were the errors below the
>>>>>> only ones you got from ansible?
>>>>>>
>>>>>>
>>>>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Wait for the host to be
>>>>>>> up]
>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts":
>>>>>>> {"ovirt_hosts": []}, "attempts": 120, "changed": false, "deprecations":
>>>>>>> [{"msg": "The 'ovirt_host_facts' module has been renamed to
>>>>>>> 'ovirt_host_info', and the renamed one no longer returns ansible_facts",
>>>>>>> "version": "2.13"}]}
>>>>>>> [ INFO  ] TASK [ovirt.hosted_engine_setup : Notify the user about a
>>>>>>> failure]
>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
>>>>>>> "The system may not be provisioned according to the playbook results:
>>>>>>> please check the logs for the issue, fix accordingly or re-deploy from
>>>>>>> scratch.\n"}
>>>>>>> [ ERROR ] Failed to execute stage 'Closing up': Failed executing
>>>>>>> ansible-playbook
>>>>>>> [ INFO  ] Stage: Termination
>>>>>>> [ ERROR ] Hosted Engine deployment failed: please check the logs for
>>>>>>> the issue, fix accordingly or re-deploy from scratch.
>>>>>>>           Log file is located at
>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20200126170315-req4qb.log
>>>>>>>
>>>>>>> *But the "real" problem seems to be SSH related, as you can see
>>>>>>> below*
>>>>>>>
>>>>>>
>>>>>> Indeed
>>>>>>
>>>>>>
>>>>>>> *[root@bric-ovirt-1 ovirt-engine]# pwd*
>>>>>>>
>>>>>>> /var/log/ovirt-hosted-engine-setup/engine-logs-2020-01-26T17:19:28Z/ovirt-engine
>>>>>>> *[root@bric-ovirt-1 ovirt-engine]# grep -i error engine.log*
>>>>>>> 2020-01-26 17:26:50,178Z ERROR
>>>>>>> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1)
>>>>>>> [2341fd23-f0c7-4f1c-ad48-88af20c2d04b] Failed to establish session with
>>>>>>> host 'bric-ovirt-1.corp.modmed.com': SSH session closed during
>>>>>>> connection '[email protected]'
>>>>>>>
>>>>>>
>>>>>> Please check/share the entire portion of engine.log, from where it
>>>>>> starts to try to ssh til it gives up.
>>>>>>
>>>>>>
>>>>>>> 2020-01-26 17:26:50,205Z ERROR
>>>>>>> [org.ovirt.engine.api.restapi.resource.AbstractBackendResource] (default
>>>>>>> task-1) [] Operation Failed: [Cannot add Host. Connecting to host via 
>>>>>>> SSH
>>>>>>> has failed, verify that the host is reachable (IP address, routable 
>>>>>>> address
>>>>>>> etc.) You may refer to the engine.log file for further details.]
>>>>>>>
>>>>>>> *The funny thing is that the engine can indeed ssh to bric-ovirt-1
>>>>>>> (physical host). See below*
>>>>>>>
>>>>>>> *[root@bric-ovirt-1 ovirt-hosted-engine-setup]# cat /etc/hosts*
>>>>>>> 192.168.1.52 bric-ovirt-engine.corp.modmed.com # temporary entry
>>>>>>> added by hosted-engine-setup for the bootstrap VM
>>>>>>> 127.0.0.1   localhost localhost.localdomain localhost4
>>>>>>> localhost4.localdomain4
>>>>>>> #::1         localhost localhost.localdomain localhost6
>>>>>>> localhost6.localdomain6
>>>>>>> 10.130.0.50 bric-ovirt-engine bric-ovirt-engine.corp.modmed.com
>>>>>>> 10.130.0.51 bric-ovirt-1 bric-ovirt-1.corp.modmed.com
>>>>>>> 10.130.0.52 bric-ovirt-2 bric-ovirt-2.corp.modmed.com
>>>>>>> 10.130.0.53 bric-ovirt-3 bric-ovirt-3.corp.modmed.com
>>>>>>> 192.168.0.1 bric-ovirt-1gluster bric-ovirt-1gluster.corp.modmed.com
>>>>>>> 192.168.0.2 bric-ovirt-2gluster bric-ovirt-2gluster.corp.modmed.com
>>>>>>> 192.168.0.3 bric-ovirt-3gluster bric-ovirt-3gluster.corp.modmed.com
>>>>>>> [root@bric-ovirt-1 ovirt-hosted-engine-setup]#
>>>>>>>
>>>>>>> *[root@bric-ovirt-1 ~]# ssh 192.168.1.52*
>>>>>>> Last login: Sun Jan 26 17:55:20 2020 from 192.168.1.1
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1*
>>>>>>> Password:
>>>>>>> Password:
>>>>>>> Last failed login: Sun Jan 26 18:17:16 UTC 2020 from 192.168.1.52 on
>>>>>>> ssh:notty
>>>>>>> There was 1 failed login attempt since the last successful login.
>>>>>>> Last login: Sun Jan 26 18:16:46 2020
>>>>>>> ###################################################################
>>>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED                #
>>>>>>> #                                                                 #
>>>>>>> # This system is the property of Modernizing Medicine, Inc.       #
>>>>>>> # It is for authorized Company business purposes only.            #
>>>>>>> # All connections are monitored and recorded.                     #
>>>>>>> # Disconnect IMMEDIATELY if you are not an authorized user!       #
>>>>>>> ###################################################################
>>>>>>> [root@bric-ovirt-1 ~]#
>>>>>>> [root@bric-ovirt-1 ~]#
>>>>>>> [root@bric-ovirt-1 ~]# exit
>>>>>>> logout
>>>>>>> Connection to bric-ovirt-1 closed.
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> *[root@bric-ovirt-engine ~]# ssh bric-ovirt-1.corp.modmed.com
>>>>>>> <http://bric-ovirt-1.corp.modmed.com>*
>>>>>>> Password:
>>>>>>> Last login: Sun Jan 26 18:17:22 2020 from 192.168.1.52
>>>>>>> ###################################################################
>>>>>>> # UNAUTHORIZED ACCESS TO THIS SYSTEM IS PROHIBITED                #
>>>>>>> #                                                                 #
>>>>>>> # This system is the property of Modernizing Medicine, Inc.       #
>>>>>>> # It is for authorized Company business purposes only.            #
>>>>>>> # All connections are monitored and recorded.                     #
>>>>>>> # Disconnect IMMEDIATELY if you are not an authorized user!       #
>>>>>>> ###################################################################
>>>>>>>
>>>>>>
>>>>>> Can you please try this, from the engine machine:
>>>>>>
>>>>>> ssh [email protected] true
>>>>>>
>>>>>> If this outputs the above "PROHIBITED" note, you'll have to configure
>>>>>> your
>>>>>> scripts etc. to not output it on non-interactive shells. Otherwise,
>>>>>> this
>>>>>> confuses the engine - it can't really distinguish between your own
>>>>>> output
>>>>>> and the output of the commands it runs there.
>>>>>>
>>>>>>
>>>>>>> [root@bric-ovirt-1 ~]# exit
>>>>>>> logout
>>>>>>> Connection to bric-ovirt-1.corp.modmed.com closed.
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> [root@bric-ovirt-engine ~]#
>>>>>>> [root@bric-ovirt-engine ~]# exit
>>>>>>> logout
>>>>>>> Connection to 192.168.1.52 closed.
>>>>>>> [root@bric-ovirt-1 ~]#
>>>>>>>
>>>>>>> *So, what gives? I already disabled all ssh security in the physical
>>>>>>> host, and whitelisted all potential IPs from the engine using firewalld.
>>>>>>> Regardless, the engine can ssh to the host as root :-(. Is there maybe
>>>>>>> another user that's used for the "Wait for the host to be  up" SSH test?
>>>>>>> Yes, I tried both passwords and certificates.*
>>>>>>>
>>>>>>
>>>>>> No, that's root. You can also see that in the log.
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> *Maybe what's really happening is that engine is not getting the
>>>>>>> right IP? bric-ovirt-engine is supposed to get 10.130.0.50, instead it
>>>>>>> never gets there, getting 192.168.1.52 from virbr0 in bric-ovirt-1. See
>>>>>>> below.*
>>>>>>>
>>>>>>
>>>>>> That's by design. For details, if interested, see "Hosted Engine 4.3
>>>>>> Deep Dive" presentation:
>>>>>>
>>>>>>
>>>>>> https://www.ovirt.org/community/get-involved/resources/slide-decks.html
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>  --== HOST NETWORK CONFIGURATION ==--
>>>>>>>           Please indicate the gateway IP address [10.130.0.1]
>>>>>>>           Please indicate a nic to set ovirtmgmt bridge on: (p4p1,
>>>>>>> p5p1) [p4p1]:
>>>>>>> --== VM CONFIGURATION ==--
>>>>>>> You may specify a unicast MAC address for the VM or accept a
>>>>>>> randomly generated default [00:16:3e:17:1d:f8]:
>>>>>>>           How should the engine VM network be configured (DHCP,
>>>>>>> Static)[DHCP]? static
>>>>>>>           Please enter the IP address to be used for the engine VM
>>>>>>> []: 10.130.0.50
>>>>>>> [ INFO  ] The engine VM will be configured to use 10.130.0.50/25
>>>>>>>           Please provide a comma-separated list (max 3) of IP
>>>>>>> addresses of domain name servers for the engine VM
>>>>>>>           Engine VM DNS (leave it empty to skip)
>>>>>>> [10.130.0.2,10.130.0.3]:
>>>>>>>           Add lines for the appliance itself and for this host to
>>>>>>> /etc/hosts on the engine VM?
>>>>>>>           Note: ensuring that this host could resolve the engine VM
>>>>>>> hostname is still up to you
>>>>>>>           (Yes, No)[No] Yes
>>>>>>>
>>>>>>> *[root@bric-ovirt-1 ~]# ip addr*
>>>>>>> 3: p4p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state
>>>>>>> UP group default qlen 1000
>>>>>>>     link/ether 00:0a:f7:f1:c6:80 brd ff:ff:ff:ff:ff:ff
>>>>>>>     inet 10.130.0.51/25 brd 10.130.0.127 scope global noprefixroute
>>>>>>> p4p1
>>>>>>>        valid_lft forever preferred_lft forever
>>>>>>> 28: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>>>>>>> state UP group default qlen 1000
>>>>>>>     link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff
>>>>>>>     inet 192.168.1.1/24 brd 192.168.1.255 scope global virbr0
>>>>>>>        valid_lft forever preferred_lft forever
>>>>>>> 29: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast
>>>>>>> master virbr0 state DOWN group default qlen 1000
>>>>>>>     link/ether 52:54:00:25:7b:6f brd ff:ff:ff:ff:ff:ff
>>>>>>> 30: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>>>> pfifo_fast master virbr0 state UNKNOWN group default qlen 1000
>>>>>>>     link/ether fe:16:3e:17:1d:f8 brd ff:ff:ff:ff:ff:ff
>>>>>>>
>>>>>>> *The newly created engine VM does remain up even after hosted-engine
>>>>>>> --deploy errors out; just at the wrong IP. I haven't been able to make 
>>>>>>> it
>>>>>>> get its real IP.*
>>>>>>>
>>>>>>
>>>>>> This happens only after the real engine VM is created, connected to
>>>>>> the correct network.
>>>>>>
>>>>>> The current engine vm you see is a libvirt VM connected to its
>>>>>> default (internal) network.
>>>>>>
>>>>>>
>>>>>>> * At any rate, thank you very much for taking a look at my very long
>>>>>>> email. Any and all help would be really appreciated.*
>>>>>>>
>>>>>>
>>>>>> Good luck and best regards,
>>>>>> --
>>>>>> Didi
>>>>>> _______________________________________________
>>>>>> Users mailing list -- [email protected]
>>>>>> To unsubscribe send an email to [email protected]
>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/AZFPSDPBK3BJUB2NESCOWQ7FQT572Y5I/
>>>>>>
>>>>>
>>>> *CONFIDENTIALITY NOTICE:* This e-mail message may contain material
>>>> protected by the Health Insurance Portability and Accountability Act of
>>>> 1996 and its implementing regulations and other state and federal laws and
>>>> legal privileges. This message is only for the personal and confidential
>>>> use of the individuals or organization to whom the message is addressed. If
>>>> you are an unintended recipient, you have received this message in error,
>>>> and any reading, distributing, copying or disclosure is unauthorized and
>>>> strictly prohibited.  All recipients are hereby notified that any
>>>> unauthorized receipt does not waive any confidentiality obligations or
>>>> privileges. If you have received this message in error, please notify the
>>>> sender immediately at the above email address and confirm that you
>>>> have deleted or destroyed the message.
>>>
>>>
>>
>>
>>
>> *CONFIDENTIALITY NOTICE:* This e-mail message may contain material
>> protected by the Health Insurance Portability and Accountability Act of
>> 1996 and its implementing regulations and other state and federal laws and
>> legal privileges. This message is only for the personal and confidential
>> use of the individuals or organization to whom the message is addressed. If
>> you are an unintended recipient, you have received this message in error,
>> and any reading, distributing, copying or disclosure is unauthorized and
>> strictly prohibited.  All recipients are hereby notified that any
>> unauthorized receipt does not waive any confidentiality obligations or
>> privileges. If you have received this message in error, please notify the
>> sender immediately at the above email address and confirm that you have
>> deleted or destroyed the message.
>
>

-- 
Didi
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/CBN2YNUCFD66SHB4BPQ2CCJFUFEJ5OOQ/

Reply via email to