On Mon, Nov 7, 2016 at 3:27 PM, Gary Pedretty <[email protected]> wrote:
> [root@fai-kvm-1-gfs admin]# gluster volume status data2 > Status of volume: data2 > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------ > ------------------ > Brick fai-kvm-1-vmn.ravnalaska.net:/kvm2/gl > uster/data2/brick 49156 0 Y > 3484 > Brick fai-kvm-2-vmn.ravnalaska.net:/kvm2/gl > uster/data2/brick 49156 0 Y > 34791 > Brick fai-kvm-3-vmn.ravnalaska.net:/kvm2/gl > uster/data2/brick 49156 0 Y > 177340 > Brick fai-kvm-4-vmn.ravnalaska.net:/kvm2/gl > uster/data2/brick 49152 0 Y > 146038 > NFS Server on localhost 2049 0 Y > 40844 > Self-heal Daemon on localhost N/A N/A Y > 40865 > NFS Server on fai-kvm-2-gfs.ravnalaska.net 2049 0 Y > 99905 > Self-heal Daemon on fai-kvm-2-gfs.ravnalask > a.net N/A N/A Y > 99915 > NFS Server on fai-kvm-4-gfs.ravnalaska.net 2049 0 Y > 176305 > Self-heal Daemon on fai-kvm-4-gfs.ravnalask > a.net N/A N/A Y > 176326 > NFS Server on fai-kvm-3-gfs.ravnalaska.net 2049 0 Y > 226271 > Self-heal Daemon on fai-kvm-3-gfs.ravnalask > a.net N/A N/A Y > 226287 > > Task Status of Volume data2 > ------------------------------------------------------------ > ------------------ > There are no active volume tasks > > > [root@fai-kvm-1-gfs admin]# gluster volume info data2 > > Volume Name: data2 > Type: Striped-Replicate > Volume ID: 20f85c9a-541b-4df4-9dba-44c5179bbfb0 > Status: Started > Number of Bricks: 1 x 2 x 2 = 4 > Transport-type: tcp > Bricks: > Brick1: fai-kvm-1-vmn.ravnalaska.net:/kvm2/gluster/data2/brick > Brick2: fai-kvm-2-vmn.ravnalaska.net:/kvm2/gluster/data2/brick > Brick3: fai-kvm-3-vmn.ravnalaska.net:/kvm2/gluster/data2/brick > Brick4: fai-kvm-4-vmn.ravnalaska.net:/kvm2/gluster/data2/brick > Options Reconfigured: > performance.readdir-ahead: on > performance.quick-read: off > performance.read-ahead: off > performance.io-cache: off > performance.stat-prefetch: off > cluster.eager-lock: enable > network.remote-dio: enable > cluster.quorum-type: auto > cluster.server-quorum-type: server > storage.owner-uid: 36 > storage.owner-gid: 36 > > > See attached file for the mount log. > Striped-Replicate is no longer supported in GlusterFS upstream. Instead, you should be using a Distribute-Replicate with sharding enabled. Also when using a gluster volume as storage domain, it is recommended to use replica 3. >From the mount logs, there is no indication as to why the volume is unmounted frequently. Could you try again with a replica 3 volume that has sharding enabled? > > Gary > > > ------------------------------------------------------------------------ > Gary Pedretty [email protected] > <[email protected]> > Systems Manager www.flyravn.com > Ravn Alaska /\ 907-450-7251 > 5245 Airport Industrial Road / \/\ 907-450-7238 fax > Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment > Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as > Really loving the record green up date! Summmer!! yourself” Matt 22:39 > ------------------------------------------------------------------------ > > > > > > > > > > > > > On Nov 6, 2016, at 9:50 PM, Sahina Bose <[email protected]> wrote: > > However your volume configuration seems suspect -"stripe 2 replica 2". Can > you provide gluster volume info of your second storage domain gluster > volume? The mount logs of the volume (under > /var/log/glusterfs/rhev-datacenter..<volname>.log) > from the host where the volume is being mounted will also help. > > > > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

