Hi Strahil Thanks for the quick answer. I will try to rsync them manually like you suggested. I am still on 4.2.x. I am in the process of moving my cluster to 4.3 but need to move to 4.2.8 first. But moving to 4.2.8 is not an easy task since I need to pin the base os to 7.6 before moving to 4.2.8. Hope moving to 4.3 will be easy :-) ... I suspect 4.4 to be a pain to upgrade since there is no upgrade path from 7.8 -> 8 ... :-( Anyway thanks for the hints.
Regards Carl On Fri, Aug 7, 2020 at 2:00 PM Strahil Nikolov <[email protected]> wrote: > I think Ravi made a change to prevent that in gluster v6.6 > > You can rsync the 2 files from ovhost1 and run a full heal (I don't > know why heal without 'full' doesn't clean up the entries). > > Anyways, ovirt can live without these 2 , but as you don't want to risk > any downtimes - just rsync them from ovhost1 and run a 'gluster volume > heal data full'. > > By the way , which version of ovirt do you use ? Gluster v3 was used in > 4.2.X > > Best Regards, > Strahil Nikolov > > > > На 7 август 2020 г. 20:14:07 GMT+03:00, carl langlois < > [email protected]> написа: > >Hi all, > > > >I am currently upgrading my ovirt cluster and after doing the upgrade > >on > >one node i end up having unsync entries that heal by the headl command. > >My setup is a 2+1 with 4 volume. > >here is a snapshot of one a volume info > >Volume Name: data > >Type: Replicate > >Volume ID: 71c999a4-b769-471f-8169-a1a66b28f9b0 > >Status: Started > >Snapshot Count: 0 > >Number of Bricks: 1 x (2 + 1) = 3 > >Transport-type: tcp > >Bricks: > >Brick1: ovhost1:/gluster_bricks/data/data > >Brick2: ovhost2:/gluster_bricks/data/data > >Brick3: ovhost3:/gluster_bricks/data/data (arbiter) > >Options Reconfigured: > >server.allow-insecure: on > >nfs.disable: on > >transport.address-family: inet > >performance.quick-read: off > >performance.read-ahead: off > >performance.io-cache: off > >performance.low-prio-threads: 32 > >network.remote-dio: enable > >cluster.eager-lock: enable > >cluster.quorum-type: auto > >cluster.server-quorum-type: server > >cluster.data-self-heal-algorithm: full > >cluster.locking-scheme: granular > >cluster.shd-max-threads: 8 > >cluster.shd-wait-qlength: 10000 > >features.shard: on > >user.cifs: off > >storage.owner-uid: 36 > >storage.owner-gid: 36 > >network.ping-timeout: 30 > >performance.strict-o-direct: on > >cluster.granular-entry-heal: enable > >features.shard-block-size: 64MB > > > >Also the output of v headl data info > > > >gluster> v heal data info > >Brick ovhost1:/gluster_bricks/data/data > >/4e59777c-5b7b-4bf1-8463-1c818067955e/dom_md/ids > >/__DIRECT_IO_TEST__ > >Status: Connected > >Number of entries: 2 > > > >Brick ovhost2:/gluster_bricks/data/data > >Status: Connected > >Number of entries: 0 > > > >Brick ovhost3:/gluster_bricks/data/data > >/4e59777c-5b7b-4bf1-8463-1c818067955e/dom_md/ids > >/__DIRECT_IO_TEST__ > >Status: Connected > >Number of entries: 2 > > > >does not seem to be a split brain also. > >gluster> v heal data info split-brain > >Brick ovhost1:/gluster_bricks/data/data > >Status: Connected > >Number of entries in split-brain: 0 > > > >Brick ovhost2:/gluster_bricks/data/data > >Status: Connected > >Number of entries in split-brain: 0 > > > >Brick ovhost3:/gluster_bricks/data/data > >Status: Connected > >Number of entries in split-brain: 0 > > > >not sure how to resolve this issue. > >gluster version is 3.2.15 > > > >Regards > > > >Carl >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
