Den 21 jul 2016 7:54 em skrev David Gossage <[email protected]>: > > On Thu, Jul 21, 2016 at 11:47 AM, Scott <[email protected]> wrote: >> >> Hi David, >> >> My backend storage is ZFS. >> >> I thought about moving from FUSE to NFS mounts for my Gluster volumes to >> help test. But since I use hosted engine this would be a real pain. Its >> difficult to modify the storage domain type/path in the hosted-engine.conf. >> And I don't want to go through the process of re-deploying hosted engine. >> > > I found this > > https://bugzilla.redhat.com/show_bug.cgi?id=1347553 > > Not sure if related. > > But I also have zfs backend, another user in gluster mailing list had issues > and used zfs backend although she used proxmox and got it working by changing > disk to writeback cache I think it was.
David and Scott, just out of curiousity, what is the OS under ZFS? /K > > I also use hosted engine, but I run my gluster volume for HE actually on a > LVM separate from zfs on xfs and if i recall it did not have the issues my > gluster on zfs did. I'm wondering now if the issue was zfs settings. > > Hopefully should have a test machone up soon I can play around with more. > >> Scott >> >> On Thu, Jul 21, 2016 at 11:36 AM David Gossage <[email protected]> >> wrote: >>> >>> What back end storage do you run gluster on? xfs/zfs/ext4 etc? >>> >>> David Gossage >>> Carousel Checks Inc. | System Administrator >>> Office 708.613.2284 >>> >>> On Thu, Jul 21, 2016 at 8:18 AM, Scott <[email protected]> wrote: >>>> >>>> I get similar problems with oVirt 4.0.1 and hosted engine. After >>>> upgrading all my hosts to Gluster 3.7.13 (client and server), I get the >>>> following: >>>> >>>> $ sudo hosted-engine --set-maintenance --mode=none >>>> Traceback (most recent call last): >>>> File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main >>>> "__main__", fname, loader, pkg_name) >>>> File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code >>>> exec code in run_globals >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>>> line 73, in <module> >>>> if not maintenance.set_mode(sys.argv[1]): >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/set_maintenance.py", >>>> line 61, in set_mode >>>> value=m_global, >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>>> line 259, in set_maintenance_mode >>>> str(value)) >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", >>>> line 204, in set_global_md_flag >>>> all_stats = broker.get_stats_from_storage(service) >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>>> line 232, in get_stats_from_storage >>>> result = self._checked_communicate(request) >>>> File >>>> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", >>>> line 260, in _checked_communicate >>>> .format(message or response)) >>>> ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request failed: failed >>>> to read metadata: [Errno 1] Operation not permitted >>>> >>>> If I only upgrade one host, then things will continue to work but my nodes >>>> are constantly healing shards. My logs are also flooded with: >>>> >>>> [2016-07-21 13:15:14.137734] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274714: READ => -1 gfid=4 >>>> 41f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not >>>> permitted) >>>> The message "W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>>> operation failed [Operation not permitted]" repeated 6 times between >>>> [2016-07-21 13:13:24.134985] and [2016-07-21 13:15:04.132226] >>>> The message "W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>>> operation failed [Operation not permitted]" repeated 8 times between >>>> [2016-07-21 13:13:34.133116] and [2016-07-21 13:15:14.137178] >>>> The message "W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>>> operation failed [Operation not permitted]" repeated 7 times between >>>> [2016-07-21 13:13:24.135071] and [2016-07-21 13:15:14.137666] >>>> [2016-07-21 13:15:24.134647] W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote >>>> operation failed [Operation not permitted] >>>> [2016-07-21 13:15:24.134764] W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote >>>> operation failed [Operation not permitted] >>>> [2016-07-21 13:15:24.134793] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274741: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation not >>>> permitted) >>>> [2016-07-21 13:15:34.135413] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274756: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not >>>> permitted) >>>> [2016-07-21 13:15:44.141062] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274818: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation not >>>> permitted) >>>> [2016-07-21 13:15:54.133582] W [MSGID: 114031] >>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote >>>> operation failed [Operation not permitted] >>>> [2016-07-21 13:15:54.133629] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274853: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation not >>>> permitted) >>>> [2016-07-21 13:16:04.133666] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274879: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not >>>> permitted) >>>> [2016-07-21 13:16:14.134954] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>> 0-glusterfs-fuse: 274894: READ => -1 >>>> gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation not >>>> permitted) >>>> >>>> Scott >>>> >>>> >>>> On Thu, Jul 21, 2016 at 6:57 AM Frank Rothenstein >>>> <[email protected]> wrote: >>>>> >>>>> Hey Devid, >>>>> >>>>> I have the very same problem on my test-cluster, despite on running ovirt >>>>> 4.0. >>>>> If you access your volumes via NFS all is fine, problem is FUSE. I stayed >>>>> on 3.7.13, but have no solution yet, now I use NFS. >>>>> >>>>> Frank >>>>> >>>>> Am Donnerstag, den 21.07.2016, 04:28 -0500 schrieb David Gossage: >>>>>> >>>>>> Anyone running one of recent 3.6.x lines and gluster using 3.7.13? I am >>>>>> looking to upgrade gluster from 3.7.11->3.7.13 for some bug fixes, but >>>>>> have been told by users on gluster mail list due to some gluster changes >>>>>> I'd need to change the disk parameters to use writeback cache. >>>>>> Something to do with aio support being removed. >>>>>> >>>>>> I believe this could be done with custom parameters? But I believe >>>>>> strage tests are done using dd and would they fail with current settings >>>>>> then? Last upgrade to 3.7.13 I had to rollback to 3.7.11 due to >>>>>> stability isues where gluster storage would go into down state and >>>>>> always show N/A as space available/used. Even if hosts saw storage >>>>>> still and VM's were running on it on all 3 hosts. >>>>>> >>>>>> Saw a lot of messages like these that went away once gluster rollback >>>>>> finished >>>>>> >>>>>> [2016-07-09 15:27:46.935694] I [fuse-bridge.c:4083:fuse_init] >>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 >>>>>> kernel 7.22 >>>>>> [2016-07-09 15:27:49.555466] W [MSGID: 114031] >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: remote >>>>>> operation failed [Operation not permitted] >>>>>> [2016-07-09 15:27:49.556574] W [MSGID: 114031] >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: remote >>>>>> operation failed [Operation not permitted] >>>>>> [2016-07-09 15:27:49.556659] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>>>> 0-glusterfs-fuse: 80: READ => -1 >>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 (Operation >>>>>> not permitted) >>>>>> [2016-07-09 15:27:59.612477] W [MSGID: 114031] >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: remote >>>>>> operation failed [Operation not permitted] >>>>>> [2016-07-09 15:27:59.613700] W [MSGID: 114031] >>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: remote >>>>>> operation failed [Operation not permitted] >>>>>> [2016-07-09 15:27:59.613781] W [fuse-bridge.c:2227:fuse_readv_cbk] >>>>>> 0-glusterfs-fuse: 168: READ => -1 >>>>>> gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 (Operation >>>>>> not permitted) >>>>>> >>>>>> David Gossage >>>>>> Carousel Checks Inc. | System Administrator >>>>>> Office 708.613.2284 >>>>>> >>>>>> _______________________________________________ >>>>>> >>>>>> Users mailing list >>>>>> >>>>>> [email protected] >>>>>> >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>> >>>>> >>>>> >>>>> ________________________________ >>>>> >>>>> >>>>> >>>>> ______________________________________________________________________________ >>>>> BODDEN-KLINIKEN Ribnitz-Damgarten GmbH >>>>> Sandhufe 2 >>>>> 18311 Ribnitz-Damgarten >>>>> >>>>> Telefon: 03821-700-0 >>>>> Fax: 03821-700-240 >>>>> >>>>> E-Mail: [email protected] Internet: http://www.bodden-kliniken.de >>>>> >>>>> Sitz: Ribnitz-Damgarten, Amtsgericht: Stralsund, HRB 2919, Steuer-Nr.: >>>>> 079/133/40188 >>>>> Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer: Dr. Falko >>>>> Milski >>>>> >>>>> Der Inhalt dieser E-Mail ist ausschließlich für den bezeichneten >>>>> Adressaten bestimmt. Wenn Sie nicht der vorge- >>>>> sehene Adressat dieser E-Mail oder dessen Vertreter sein sollten, >>>>> beachten Sie bitte, dass jede Form der Veröf- >>>>> fentlichung, Vervielfältigung oder Weitergabe des Inhalts dieser E-Mail >>>>> unzulässig ist. Wir bitten Sie, sofort den >>>>> Absender zu informieren und die E-Mail zu löschen. >>>>> >>>>> >>>>> Bodden-Kliniken Ribnitz-Damgarten GmbH 2016 >>>>> *** Virenfrei durch Kerio Mail Server und Sophos Antivirus *** >>>>> _______________________________________________ >>>>> Users mailing list >>>>> [email protected] >>>>> http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

