On 30 July 2016 at 02:48, Kenneth Bingham <[email protected]> wrote: > Aw crap. I did exactly the same thing and this could explain a lot of the > issues I've been pulling out my beard over. Every time I did 'hosted-engine > --deploy' on the RHEV-M|NODE host I entered the FQDN of *that* host, not > the first host, as the origin of the Gluster FS volume because at the time > I didn't realize that > a. the manager would key deduplication on the URI of the volume > b. that the volume would be mounted on FUSE, not NFS, and therefore no > single point of entry is created by using the FQDN of the first host > because the GFS client will persist connections with all peers > > If you ever want to add an hosted engine host to your setup please do that from UI and not from CLI. That will prevent all this confusion.
> On Fri, Jul 29, 2016 at 6:08 AM Simone Tiraboschi <[email protected]> > wrote: > >> On Fri, Jul 29, 2016 at 11:35 AM, Wee Sritippho <[email protected]> >> wrote: >> > On 29/7/2559 15:50, Simone Tiraboschi wrote: >> >> >> >> On Fri, Jul 29, 2016 at 6:31 AM, Wee Sritippho <[email protected]> >> wrote: >> >>> >> >>> On 28/7/2559 15:54, Simone Tiraboschi wrote: >> >>> >> >>> On Thu, Jul 28, 2016 at 10:41 AM, Wee Sritippho <[email protected]> >> >>> wrote: >> >>>> >> >>>> On 21/7/2559 16:53, Simone Tiraboschi wrote: >> >>>> >> >>>> On Thu, Jul 21, 2016 at 11:43 AM, Wee Sritippho <[email protected]> >> >>>> wrote: >> >>>> >> >>>>> Can I just follow >> >>>>> >> >>>>> >> http://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >> >>>>> until step 3 and do everything else via GUI? >> >>>> >> >>>> Yes, absolutely. >> >>>> >> >>>> >> >>>> Hi, I upgrade a host (host02) via GUI and now its score is 0. >> Restarted >> >>>> the services but the result is still the same. Kinda lost now. What >> >>>> should I >> >>>> do next? >> >>>> >> >>> Can you please attach ovirt-ha-agent logs? >> >>> >> >>> >> >>> Yes, here are the logs: >> >>> https://app.box.com/s/b4urjty8dsuj98n3ywygpk3oh5o7pbsh >> >> >> >> Thanks Wee, >> >> your issue is here: >> >> MainThread::ERROR::2016-07-17 >> >> >> >> >> 14:32:45,586::storage_server::143::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(_validate_pre_connected_path) >> >> The hosted-engine storage domain is already mounted on >> >> >> >> '/rhev/data-center/mnt/glusterSD/host02.ovirt.forest.go.th: >> _hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4' >> >> with a path that is not supported anymore: the right path should be >> >> >> >> '/rhev/data-center/mnt/glusterSD/host01.ovirt.forest.go.th: >> _hosted__engine/639e689c-8493-479b-a6eb-cc92b6fc4cf4'. >> >> >> >> Did you manually tried to avoid the issue of a single entry point for >> >> the gluster FS volume using host01.ovirt.forest.go.th:_hosted__engine >> >> and host02.ovirt.forest.go.th:_hosted__engine there? >> >> This could cause a lot of confusion since the code could not detect >> >> that the storage domain is the same and you can end with it mounted >> >> twice into different locations and a lot of issues. >> >> The correct solution of that issue was this one: >> >> https://bugzilla.redhat.com/show_bug.cgi?id=1298693#c20 >> >> >> >> Now, to have it fixed on your env you have to hack a bit. >> >> First step, you have to edit >> >> /etc/ovirt-hosted-engine/hosted-engine.conf on all your hosted-engine >> >> hosts to ensure that the storage field always point to the same entry >> >> point (host01 for instance) >> >> Then on each host you can add something like: >> >> >> >> mnt_options=backupvolfile-server=host02.ovirt.forest.go.th:h >> ost03.ovirt.forest.go.th >> ,fetch-attempts=2,log-level=WARNING,log-file=/var/log/engine_domain.log >> >> >> >> Then check the representation of your storage connection in the table >> >> storage_server_connections of the engine DB and make sure that >> >> connection refers to the entry point you used in hosted-engine.conf on >> >> all your hosts, you have lastly to set the value of mount_options also >> >> here. >> > >> > Weird. The configuration in all hosts are already referring to host01. >> >> but for sure you have a connection pointing to host02 somewhere, did >> you try to manually deploy from CLI connecting the gluster volume on >> host02? >> >> > Also, in the storage_server_connections table: >> > >> > engine=> SELECT * FROM storage_server_connections; >> > id | connection | >> > user_name | password | iqn | port | portal | storage_type | >> mount_options | >> > vfs_type >> > | nfs_version | nfs_timeo | nfs_retrans >> > >> --------------------------------------+------------------------------------------+-----------+----------+-----+------+--------+--------------+---------------+---------- >> > -+-------------+-----------+------------- >> > bd78d299-c8ff-4251-8aab-432ce6443ae8 | >> > host01.ovirt.forest.go.th:/hosted_engine | | | | | 1 >> > | 7 | | glusterfs >> > | | | >> > (1 row) >> > >> > >> >> >> >> Please tune also the value of network.ping-timeout for your glusterFS >> >> volume to avoid this: >> >> https://bugzilla.redhat.com/show_bug.cgi?id=1319657#c17 >> > >> > >> > -- >> > Wee >> > >> _______________________________________________ >> Users mailing list >> [email protected] >> http://lists.ovirt.org/mailman/listinfo/users >> > > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

