Nope, not in any official repo. I only use those suggested by oVirt, ie: http://centos.bhs.mirrors.ovh.net/ftp.centos.org/7/storage/x86_64/gluster-3.7/
No 3.7.14 there. Thanks though. Scott On Sat, Aug 13, 2016 at 11:23 AM David Gossage <[email protected]> wrote: > On Sat, Aug 13, 2016 at 11:00 AM, David Gossage < > [email protected]> wrote: > >> On Sat, Aug 13, 2016 at 8:19 AM, Scott <[email protected]> wrote: >> >>> Had a chance to upgrade my cluster to Gluster 3.7.14 and can confirm it >>> works for me too where 3.7.12/13 did not. >>> >>> I did find that you should NOT turn off network.remote-dio or turn >>> on performance.strict-o-direct as suggested earlier in the thread. They >>> will prevent dd (using direct flag) and other things from working >>> properly. I'd leave those at network.remote-dio=enabled >>> and performance.strict-o-direct=off. >>> >> >> Those were actually just suggested during a testing phase trying to trace >> down the issue. Neither of those 2 I think have ever been suggested as >> good practice. At least not for VM storage. >> >> >>> Hopefully we can see Gluster 3.7.14 moved out of testing repo soon. >>> >> > Is it still in testing repo? I updated my production cluster I think 2 > weeks ago from default repo on centos7. > > >>> Scott >>> >>> On Tue, Aug 2, 2016 at 9:05 AM, David Gossage < >>> [email protected]> wrote: >>> >>>> So far gluster 3.7.14 seems to have resolved issues at least on my test >>>> box. dd commands that failed previously now work with sharding on zfs >>>> backend, >>>> >>>> Where before I couldn't even mount a new storage domain it now mounted >>>> and I have a test vm being created. >>>> >>>> Still have to let VM run for a few days and make sure no locking >>>> freezing occurs but looks hopeful so far. >>>> >>>> *David Gossage* >>>> *Carousel Checks Inc. | System Administrator* >>>> *Office* 708.613.2284 >>>> >>>> On Tue, Jul 26, 2016 at 8:15 AM, David Gossage < >>>> [email protected]> wrote: >>>> >>>>> On Tue, Jul 26, 2016 at 4:37 AM, Krutika Dhananjay < >>>>> [email protected]> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> 1. Could you please attach the glustershd logs from all three nodes? >>>>>> >>>>> >>>>> Here are ccgl1 and ccgl2. as previously mentioned ccgl3 third node >>>>> was down from bad nic so no relevant logs would be on that node. >>>>> >>>>> >>>>>> >>>>>> 2. Also, so far what we know is that the 'Operation not permitted' >>>>>> errors are on the main vm image itself and not its individual shards (ex >>>>>> deb61291-5176-4b81-8315-3f1cf8e3534d). Could you do the following: >>>>>> Get the inode number of >>>>>> .glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d (ls -li) from the >>>>>> first brick. I'll call this number INODE_NUMBER. >>>>>> Execute `find . -inum INODE_NUMBER` from the brick root on first >>>>>> brick to print the hard links against the file in the prev step and share >>>>>> the output. >>>>>> >>>>> [dgossage@ccgl1 ~]$ sudo ls -li >>>>> /gluster1/BRICK1/1/.glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d >>>>> 16407 -rw-r--r--. 2 36 36 466 Jun 5 16:52 >>>>> /gluster1/BRICK1/1/.glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d >>>>> [dgossage@ccgl1 ~]$ cd /gluster1/BRICK1/1/ >>>>> [dgossage@ccgl1 1]$ sudo find . -inum 16407 >>>>> ./7c73a8dd-a72e-4556-ac88-7f6813131e64/dom_md/metadata >>>>> ./.glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d >>>>> >>>>> >>>>> >>>>>> >>>>>> 3. Did you delete any vms at any point before or after the upgrade? >>>>>> >>>>> >>>>> Immediately before or after on same day pretty sure I deleted >>>>> nothing. During week prior I deleted a few dev vm's that were never setup >>>>> and some the week after upgrade as I was preparing for moving disks off >>>>> and >>>>> on storage to get them sharded and felt it would be easier to just >>>>> recreate >>>>> some disks that had no data yet rather than move them off and on later. >>>>> >>>>>> >>>>>> -Krutika >>>>>> >>>>>> On Mon, Jul 25, 2016 at 11:30 PM, David Gossage < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> >>>>>>> On Mon, Jul 25, 2016 at 9:58 AM, Krutika Dhananjay < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> OK, could you try the following: >>>>>>>> >>>>>>>> i. Set network.remote-dio to off >>>>>>>> # gluster volume set <VOL> network.remote-dio off >>>>>>>> >>>>>>>> ii. Set performance.strict-o-direct to on >>>>>>>> # gluster volume set <VOL> performance.strict-o-direct on >>>>>>>> >>>>>>>> iii. Stop the affected vm(s) and start again >>>>>>>> >>>>>>>> and tell me if you notice any improvement? >>>>>>>> >>>>>>>> >>>>>>> Previous instll I had issue with is still on gluster 3.7.11 >>>>>>> >>>>>>> My test install of ovirt 3.6.7 and gluster 3.7.13 with 3 bricks on a >>>>>>> locak disk right now isn't allowing me to add the gluster storage at >>>>>>> all. >>>>>>> >>>>>>> Keep getting some type of UI error >>>>>>> >>>>>>> 2016-07-25 12:49:09,277 ERROR >>>>>>> [org.ovirt.engine.ui.frontend.server.gwt.OvirtRemoteLoggingService] >>>>>>> (default task-33) [] Permutation name: 430985F23DFC1C8BE1C7FDD91EDAA785 >>>>>>> 2016-07-25 12:49:09,277 ERROR >>>>>>> [org.ovirt.engine.ui.frontend.server.gwt.OvirtRemoteLoggingService] >>>>>>> (default task-33) [] Uncaught exception: : java.lang.ClassCastException >>>>>>> at Unknown.ps( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@3837) >>>>>>> at Unknown.ts( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@20) >>>>>>> at Unknown.vs( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@18) >>>>>>> at Unknown.iJf( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@19) >>>>>>> at Unknown.Xab( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@48) >>>>>>> at Unknown.P8o( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@4447) >>>>>>> at Unknown.jQr( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@21) >>>>>>> at Unknown.A8o( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@51) >>>>>>> at Unknown.u8o( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@101) >>>>>>> at Unknown.Eap( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@10718) >>>>>>> at Unknown.p8n( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@161) >>>>>>> at Unknown.Cao( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@31) >>>>>>> at Unknown.Bap( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@10469) >>>>>>> at Unknown.kRn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@49) >>>>>>> at Unknown.nRn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@438) >>>>>>> at Unknown.eVn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@40) >>>>>>> at Unknown.hVn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@25827) >>>>>>> at Unknown.MTn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@25) >>>>>>> at Unknown.PTn( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@24052) >>>>>>> at Unknown.KJe( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@21125) >>>>>>> at Unknown.Izk( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@10384) >>>>>>> at Unknown.P3( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@137) >>>>>>> at Unknown.g4( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@8271) >>>>>>> at Unknown.<anonymous>( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@65) >>>>>>> at Unknown._t( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@29) >>>>>>> at Unknown.du( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@57) >>>>>>> at Unknown.<anonymous>( >>>>>>> https://ccengine2.carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@54 >>>>>>> ) >>>>>>> >>>>>>> >>>>>>>> -Krutika >>>>>>>> >>>>>>>> On Mon, Jul 25, 2016 at 4:57 PM, Samuli Heinonen < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> > On 25 Jul 2016, at 12:34, David Gossage < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > >>>>>>>>> > On Mon, Jul 25, 2016 at 1:01 AM, Krutika Dhananjay < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > Hi, >>>>>>>>> > >>>>>>>>> > Thanks for the logs. So I have identified one issue from the >>>>>>>>> logs for which the fix is this: >>>>>>>>> http://review.gluster.org/#/c/14669/. Because of a bug in the >>>>>>>>> code, ENOENT was getting converted to EPERM and being propagated up >>>>>>>>> the >>>>>>>>> stack causing the reads to bail out early with 'Operation not >>>>>>>>> permitted' >>>>>>>>> errors. >>>>>>>>> > I still need to find out two things: >>>>>>>>> > i) why there was a readv() sent on a non-existent (ENOENT) file >>>>>>>>> (this is important since some of the other users have not faced or >>>>>>>>> reported >>>>>>>>> this issue on gluster-users with 3.7.13) >>>>>>>>> > ii) need to see if there's a way to work around this issue. >>>>>>>>> > >>>>>>>>> > Do you mind sharing the steps needed to be executed to run into >>>>>>>>> this issue? This is so that we can apply our patches, test and ensure >>>>>>>>> they >>>>>>>>> fix the problem. >>>>>>>>> >>>>>>>>> >>>>>>>>> Unfortunately I can’t test this right away nor give exact steps >>>>>>>>> how to test this. This is just a theory but please correct me if you >>>>>>>>> see >>>>>>>>> some mistakes. >>>>>>>>> >>>>>>>>> oVirt uses cache=none settings for VM’s by default which requires >>>>>>>>> direct I/O. oVirt also uses dd with iflag=direct to check that >>>>>>>>> storage has >>>>>>>>> direct I/O enabled. Problems exist with GlusterFS with sharding >>>>>>>>> enabled and >>>>>>>>> bricks running on ZFS on Linux. Everything seems to be fine with >>>>>>>>> GlusterFS >>>>>>>>> 3.7.11 and problems exist at least with version .12 and .13. There >>>>>>>>> has been >>>>>>>>> some posts saying that GlusterFS 3.8.x is also affected. >>>>>>>>> >>>>>>>>> Steps to reproduce: >>>>>>>>> 1. Sharded file is created with GlusterFS 3.7.11. Everything works >>>>>>>>> ok. >>>>>>>>> 2. GlusterFS is upgraded to 3.7.12+ >>>>>>>>> 3. Sharded file cannot be read or written with direct I/O enabled. >>>>>>>>> (Ie. oVirt uses to check storage connection with command "dd >>>>>>>>> if=/rhev/data-center/00000001-0001-0001-0001-0000000002b6/mastersd/dom_md/inbox >>>>>>>>> iflag=direct,fullblock count=1 bs=1024000”) >>>>>>>>> >>>>>>>>> Please let me know if you need more information. >>>>>>>>> >>>>>>>>> -samuli >>>>>>>>> >>>>>>>>> > Well after upgrade of gluster all I did was start ovirt hosts up >>>>>>>>> which launched and started their ha-agent and broker processes. I >>>>>>>>> don't >>>>>>>>> believe I started getting any errors till it mounted GLUSTER1. I had >>>>>>>>> enabled sharding but had no sharded disk images yet. Not sure if the >>>>>>>>> check >>>>>>>>> for shards would have caused that. Unfortunately I can't just update >>>>>>>>> this >>>>>>>>> cluster and try and see what caused it as it has sme VM's users >>>>>>>>> expect to >>>>>>>>> be available in few hours. >>>>>>>>> > >>>>>>>>> > I can see if I can get my test setup to recreate it. I think >>>>>>>>> I'll need to de-activate data center so I can detach the storage >>>>>>>>> thats on >>>>>>>>> xfs and attach the one thats over zfs with sharding enabled. My test >>>>>>>>> is 3 >>>>>>>>> bricks on same local machine, with 3 different volumes but I think im >>>>>>>>> running into sanlock issue or something as it won't mount more than >>>>>>>>> one >>>>>>>>> volume that was created locally. >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > -Krutika >>>>>>>>> > >>>>>>>>> > On Fri, Jul 22, 2016 at 7:17 PM, David Gossage < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > Trimmed out the logs to just about when I was shutting down >>>>>>>>> ovirt servers for updates which was 14:30 UTC 2016-07-09 >>>>>>>>> > >>>>>>>>> > Pre-update settings were >>>>>>>>> > >>>>>>>>> > Volume Name: GLUSTER1 >>>>>>>>> > Type: Replicate >>>>>>>>> > Volume ID: 167b8e57-28c3-447a-95cc-8410cbdf3f7f >>>>>>>>> > Status: Started >>>>>>>>> > Number of Bricks: 1 x 3 = 3 >>>>>>>>> > Transport-type: tcp >>>>>>>>> > Bricks: >>>>>>>>> > Brick1: ccgl1.gl.local:/gluster1/BRICK1/1 >>>>>>>>> > Brick2: ccgl2.gl.local:/gluster1/BRICK1/1 >>>>>>>>> > Brick3: ccgl3.gl.local:/gluster1/BRICK1/1 >>>>>>>>> > Options Reconfigured: >>>>>>>>> > performance.readdir-ahead: on >>>>>>>>> > storage.owner-uid: 36 >>>>>>>>> > storage.owner-gid: 36 >>>>>>>>> > performance.quick-read: off >>>>>>>>> > performance.read-ahead: off >>>>>>>>> > performance.io-cache: off >>>>>>>>> > performance.stat-prefetch: off >>>>>>>>> > cluster.eager-lock: enable >>>>>>>>> > network.remote-dio: enable >>>>>>>>> > cluster.quorum-type: auto >>>>>>>>> > cluster.server-quorum-type: server >>>>>>>>> > server.allow-insecure: on >>>>>>>>> > cluster.self-heal-window-size: 1024 >>>>>>>>> > cluster.background-self-heal-count: 16 >>>>>>>>> > performance.strict-write-ordering: off >>>>>>>>> > nfs.disable: on >>>>>>>>> > nfs.addr-namelookup: off >>>>>>>>> > nfs.enable-ino32: off >>>>>>>>> > >>>>>>>>> > At the time of updates ccgl3 was offline from bad nic on server >>>>>>>>> but had been so for about a week with no issues in volume >>>>>>>>> > >>>>>>>>> > Shortly after update I added these settings to enable sharding >>>>>>>>> but did not as of yet have any VM images sharded. >>>>>>>>> > features.shard-block-size: 64MB >>>>>>>>> > features.shard: on >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > David Gossage >>>>>>>>> > Carousel Checks Inc. | System Administrator >>>>>>>>> > Office 708.613.2284 >>>>>>>>> > >>>>>>>>> > On Fri, Jul 22, 2016 at 5:00 AM, Krutika Dhananjay < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > Hi David, >>>>>>>>> > >>>>>>>>> > Could you also share the brick logs from the affected volume? >>>>>>>>> They're located at >>>>>>>>> /var/log/glusterfs/bricks/<hyphenated-path-to-the-brick-directory>.log. >>>>>>>>> > >>>>>>>>> > Also, could you share the volume configuration (output of >>>>>>>>> `gluster volume info <VOL>`) for the affected volume(s) AND at the >>>>>>>>> time you >>>>>>>>> actually saw this issue? >>>>>>>>> > >>>>>>>>> > -Krutika >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > On Thu, Jul 21, 2016 at 11:23 PM, David Gossage < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > On Thu, Jul 21, 2016 at 11:47 AM, Scott <[email protected]> >>>>>>>>> wrote: >>>>>>>>> > Hi David, >>>>>>>>> > >>>>>>>>> > My backend storage is ZFS. >>>>>>>>> > >>>>>>>>> > I thought about moving from FUSE to NFS mounts for my Gluster >>>>>>>>> volumes to help test. But since I use hosted engine this would be a >>>>>>>>> real >>>>>>>>> pain. Its difficult to modify the storage domain type/path in the >>>>>>>>> hosted-engine.conf. And I don't want to go through the process of >>>>>>>>> re-deploying hosted engine. >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > I found this >>>>>>>>> > >>>>>>>>> > https://bugzilla.redhat.com/show_bug.cgi?id=1347553 >>>>>>>>> > >>>>>>>>> > Not sure if related. >>>>>>>>> > >>>>>>>>> > But I also have zfs backend, another user in gluster mailing >>>>>>>>> list had issues and used zfs backend although she used proxmox and >>>>>>>>> got it >>>>>>>>> working by changing disk to writeback cache I think it was. >>>>>>>>> > >>>>>>>>> > I also use hosted engine, but I run my gluster volume for HE >>>>>>>>> actually on a LVM separate from zfs on xfs and if i recall it did not >>>>>>>>> have >>>>>>>>> the issues my gluster on zfs did. I'm wondering now if the issue was >>>>>>>>> zfs >>>>>>>>> settings. >>>>>>>>> > >>>>>>>>> > Hopefully should have a test machone up soon I can play around >>>>>>>>> with more. >>>>>>>>> > >>>>>>>>> > Scott >>>>>>>>> > >>>>>>>>> > On Thu, Jul 21, 2016 at 11:36 AM David Gossage < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > What back end storage do you run gluster on? xfs/zfs/ext4 etc? >>>>>>>>> > >>>>>>>>> > David Gossage >>>>>>>>> > Carousel Checks Inc. | System Administrator >>>>>>>>> > Office 708.613.2284 >>>>>>>>> > >>>>>>>>> > On Thu, Jul 21, 2016 at 8:18 AM, Scott <[email protected]> >>>>>>>>> wrote: >>>>>>>>> > I get similar problems with oVirt 4.0.1 and hosted engine. >>>>>>>>> After upgrading all my hosts to Gluster 3.7.13 (client and server), I >>>>>>>>> get >>>>>>>>> the following: >>>>>>>>> > >>>>>>>>> > $ sudo hosted-engine --set-maintenance --mode=none >>>>>>>>> > Traceback (most recent call last): >>>>>>>>> > File "/usr/lib64/python2.7/runpy.py", line 162, in >>>>>>>>> _run_module_as_main >>>>>>>>> > "__main__", fname, loader, pkg_name) >>>>>>>>> > File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code >>>>>>>>> > exec code in run_globals >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>>>>>>>> line 73, in <module> >>>>>>>>> > if not maintenance.set_mode(sys.argv[1]): >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>>>>>>>> line 61, in set_mode >>>>>>>>> > value=m_global, >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>>>>>>>> line 259, in set_maintenance_mode >>>>>>>>> > str(value)) >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>>>>>>>> line 204, in set_global_md_flag >>>>>>>>> > all_stats = broker.get_stats_from_storage(service) >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>>>>>>>> line 232, in get_stats_from_storage >>>>>>>>> > result = self._checked_communicate(request) >>>>>>>>> > File >>>>>>>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>>>>>>>> line 260, in _checked_communicate >>>>>>>>> > .format(message or response)) >>>>>>>>> > ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request >>>>>>>>> failed: failed to read metadata: [Errno 1] Operation not permitted >>>>>>>>> > >>>>>>>>> > If I only upgrade one host, then things will continue to work >>>>>>>>> but my nodes are constantly healing shards. My logs are also flooded >>>>>>>>> with: >>>>>>>>> > >>>>>>>>> > [2016-07-21 13:15:14.137734] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274714: READ => >>>>>>>>> -1 >>>>>>>>> gfid=4 >>>>>>>>> > 41f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation >>>>>>>>> not permitted) >>>>>>>>> > The message "W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>>>>>>>> operation failed [Operation not permitted]" repeated 6 times between >>>>>>>>> [2016-07-21 13:13:24.134985] and [2016-07-21 13:15:04.132226] >>>>>>>>> > The message "W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>>>>>>>> operation failed [Operation not permitted]" repeated 8 times between >>>>>>>>> [2016-07-21 13:13:34.133116] and [2016-07-21 13:15:14.137178] >>>>>>>>> > The message "W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>>>>>>>> operation failed [Operation not permitted]" repeated 7 times between >>>>>>>>> [2016-07-21 13:13:24.135071] and [2016-07-21 13:15:14.137666] >>>>>>>>> > [2016-07-21 13:15:24.134647] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> > [2016-07-21 13:15:24.134764] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> > [2016-07-21 13:15:24.134793] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274741: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > [2016-07-21 13:15:34.135413] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274756: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > [2016-07-21 13:15:44.141062] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274818: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > [2016-07-21 13:15:54.133582] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> > [2016-07-21 13:15:54.133629] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274853: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > [2016-07-21 13:16:04.133666] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274879: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > [2016-07-21 13:16:14.134954] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274894: READ => >>>>>>>>> -1 >>>>>>>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> > >>>>>>>>> > Scott >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > On Thu, Jul 21, 2016 at 6:57 AM Frank Rothenstein < >>>>>>>>> [email protected]> wrote: >>>>>>>>> > Hey Devid, >>>>>>>>> > >>>>>>>>> > I have the very same problem on my test-cluster, despite on >>>>>>>>> running ovirt 4.0. >>>>>>>>> > If you access your volumes via NFS all is fine, problem is FUSE. >>>>>>>>> I stayed on 3.7.13, but have no solution yet, now I use NFS. >>>>>>>>> > >>>>>>>>> > Frank >>>>>>>>> > >>>>>>>>> > Am Donnerstag, den 21.07.2016, 04:28 -0500 schrieb David Gossage: >>>>>>>>> >> Anyone running one of recent 3.6.x lines and gluster using >>>>>>>>> 3.7.13? I am looking to upgrade gluster from 3.7.11->3.7.13 for some >>>>>>>>> bug >>>>>>>>> fixes, but have been told by users on gluster mail list due to some >>>>>>>>> gluster >>>>>>>>> changes I'd need to change the disk parameters to use writeback cache. >>>>>>>>> Something to do with aio support being removed. >>>>>>>>> >> >>>>>>>>> >> I believe this could be done with custom parameters? But I >>>>>>>>> believe strage tests are done using dd and would they fail with >>>>>>>>> current >>>>>>>>> settings then? Last upgrade to 3.7.13 I had to rollback to 3.7.11 due >>>>>>>>> to >>>>>>>>> stability isues where gluster storage would go into down state and >>>>>>>>> always >>>>>>>>> show N/A as space available/used. Even if hosts saw storage still >>>>>>>>> and VM's >>>>>>>>> were running on it on all 3 hosts. >>>>>>>>> >> >>>>>>>>> >> Saw a lot of messages like these that went away once gluster >>>>>>>>> rollback finished >>>>>>>>> >> >>>>>>>>> >> [2016-07-09 15:27:46.935694] I [fuse-bridge.c:4083:fuse_init] >>>>>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 >>>>>>>>> kernel >>>>>>>>> 7.22 >>>>>>>>> >> [2016-07-09 15:27:49.555466] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: >>>>>>>>> remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> >> [2016-07-09 15:27:49.556574] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: >>>>>>>>> remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> >> [2016-07-09 15:27:49.556659] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 80: READ => -1 >>>>>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> >> [2016-07-09 15:27:59.612477] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: >>>>>>>>> remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> >> [2016-07-09 15:27:59.613700] W [MSGID: 114031] >>>>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: >>>>>>>>> remote >>>>>>>>> operation failed [Operation not permitted] >>>>>>>>> >> [2016-07-09 15:27:59.613781] W >>>>>>>>> [fuse-bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 168: READ => -1 >>>>>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 >>>>>>>>> (Operation not >>>>>>>>> permitted) >>>>>>>>> >> >>>>>>>>> >> David Gossage >>>>>>>>> >> Carousel Checks Inc. | System Administrator >>>>>>>>> >> Office 708.613.2284 >>>>>>>>> >> _______________________________________________ >>>>>>>>> >> Users mailing list >>>>>>>>> >> >>>>>>>>> >> [email protected] >>>>>>>>> >> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> ______________________________________________________________________________ >>>>>>>>> > BODDEN-KLINIKEN Ribnitz-Damgarten GmbH >>>>>>>>> > Sandhufe 2 >>>>>>>>> > 18311 Ribnitz-Damgarten >>>>>>>>> > >>>>>>>>> > Telefon: 03821-700-0 >>>>>>>>> > Fax: 03821-700-240 >>>>>>>>> > >>>>>>>>> > E-Mail: [email protected] Internet: >>>>>>>>> http://www.bodden-kliniken.de >>>>>>>>> > >>>>>>>>> > Sitz: Ribnitz-Damgarten, Amtsgericht: Stralsund, HRB 2919, >>>>>>>>> Steuer-Nr.: 079/133/40188 >>>>>>>>> > Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer: Dr. >>>>>>>>> Falko Milski >>>>>>>>> > >>>>>>>>> > Der Inhalt dieser E-Mail ist ausschließlich für den bezeichneten >>>>>>>>> Adressaten bestimmt. Wenn Sie nicht der vorge- >>>>>>>>> > sehene Adressat dieser E-Mail oder dessen Vertreter sein >>>>>>>>> sollten, beachten Sie bitte, dass jede Form der Veröf- >>>>>>>>> > fentlichung, Vervielfältigung oder Weitergabe des Inhalts dieser >>>>>>>>> E-Mail unzulässig ist. Wir bitten Sie, sofort den >>>>>>>>> > Absender zu informieren und die E-Mail zu löschen. >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > Bodden-Kliniken Ribnitz-Damgarten GmbH 2016 >>>>>>>>> > *** Virenfrei durch Kerio Mail Server und Sophos Antivirus *** >>>>>>>>> > _______________________________________________ >>>>>>>>> > Users mailing list >>>>>>>>> > [email protected] >>>>>>>>> > http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > _______________________________________________ >>>>>>>>> > Users mailing list >>>>>>>>> > [email protected] >>>>>>>>> > http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > _______________________________________________ >>>>>>>>> > Users mailing list >>>>>>>>> > [email protected] >>>>>>>>> > http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> [email protected] >>>> http://lists.ovirt.org/mailman/listinfo/users >>>> >>>> >>> >>
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

