Dear Sahina, Is there any changes with this bug?
Still I haven't finish my upgrade process that i've started on 9th may:( Please help me if you can. Thanks Tibor ----- 2018. máj.. 18., 9:29, Demeter Tibor <[email protected]> írta: > Hi, > I have to update the engine again? > Thanks, > R > Tibor > ----- 2018. máj.. 18., 6:47, Sahina Bose <[email protected]> írta: >> Thanks for reporting this. [ https://gerrit.ovirt.org/91375 | >> https://gerrit.ovirt.org/91375 ] fixes this. I've re-opened bug [ >> https://bugzilla.redhat.com/show_bug.cgi?id=1574508 | >> https://bugzilla.redhat.com/show_bug.cgi?id=1574508 ] >> On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor < [ >> mailto:[email protected] | >> [email protected] ] > wrote: >>> Hi, >>> 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos >>> Firstly, I did a yum update "ovirt-*-setup*" >>> second, I have ran engine-setup to upgrade. >>> I didn't remove the old repos, just installed the nightly repo. >>> Thank you again, >>> Regards, >>> Tibor >>> ----- 2018. máj.. 17., 15:02, Sahina Bose < [ mailto:[email protected] | >>> [email protected] ] > írta: >>>> It doesn't look like the patch was applied. Still see the same error in >>>> engine.log >>>> "Error while refreshing brick statuses for volume 'volume1' of cluster >>>> 'C6220': >>>> null"\ >>>> Did you use engine-setup to upgrade? What's the version of ovirt-engine >>>> currently installed? >>>> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [ >>>> mailto:[email protected] | >>>> [email protected] ] > wrote: >>>>> Hi, >>>>> sure, >>>>> Thank you for your time! >>>>> R >>>>> Tibor >>>>> ----- 2018. máj.. 17., 12:19, Sahina Bose < [ mailto:[email protected] | >>>>> [email protected] ] > írta: >>>>>> [+users] >>>>>> Can you provide the engine.log to see why the monitoring is not working >>>>>> here. >>>>>> thanks! >>>>>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [ >>>>>> mailto:[email protected] | >>>>>> [email protected] ] > wrote: >>>>>>> Hi, >>>>>>> Meanwhile, I did the upgrade engine, but the gluster state is same on >>>>>>> my first >>>>>>> node. >>>>>>> I've attached some screenshot of my problem. >>>>>>> Thanks >>>>>>> Tibor >>>>>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor < [ >>>>>>> mailto:[email protected] | >>>>>>> [email protected] ] > írta Hi, >>>>>>>> If 4.3.4 will release, i just have to remove the nightly repo and >>>>>>>> update to >>>>>>>> stable? >>>>>>>> I'm sorry for my terrible English, I try to explain what was my >>>>>>>> problem with >>>>>>>> update. >>>>>>>> I'm upgraded from 4.1.8. >>>>>>>> I followed up the official hosted-engine update documentation, that >>>>>>>> was not >>>>>>>> clear me, because it has referenced to a lot of old thing (i think). >>>>>>>> [ https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ | >>>>>>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ ] >>>>>>>> [ >>>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>>>>> | >>>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>>>>> ] >>>>>>>> Maybe it need to update, because I had a lot of question under upgrade >>>>>>>> and I was >>>>>>>> not sure in all of necessary steps. For example, If I need to >>>>>>>> installing the >>>>>>>> new, 4.2 repo on the hosts, then need to remove the old repo from that? >>>>>>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an >>>>>>>> "Updatehost" >>>>>>>> menu in the GUI? So, maybe it outdated. >>>>>>>> Since upgrade hosted engine, and the first node, I have problems with >>>>>>>> gluster. >>>>>>>> It seems to working fine if you check it from console "gluster volume >>>>>>>> status, >>>>>>>> etc" but not on the Gui, because now it yellow, and the brick reds in >>>>>>>> the first >>>>>>>> node. >>>>>>>> Previously I did a mistake with glusterfs, my gluster config was >>>>>>>> wrong. I have >>>>>>>> corrected them, but it did not helped to me,gluster bricks are reds on >>>>>>>> my first >>>>>>>> node yet.... >>>>>>>> Now I try to upgrade to nightly, but I'm affraid, because it a living, >>>>>>>> productive system, and I don't have downtime. I hope it will help me. >>>>>>>> Thanks for all, >>>>>>>> Regards, >>>>>>>> Tibor Demeter >>>>>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose < [ mailto:[email protected] | >>>>>>>> [email protected] ] > írta: >>>>>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor < [ >>>>>>>>> mailto:[email protected] | >>>>>>>>> [email protected] ] > wrote: >>>>>>>>>> Hi, >>>>>>>>>> is it a different, unstable repo? I have a productive cluster, how >>>>>>>>>> is safe that? >>>>>>>>>> I don't have any experience with nightly build. How can I use this? >>>>>>>>>> It have to >>>>>>>>>> install to the engine VM or all of my hosts? >>>>>>>>>> Thanks in advance for help me.. >>>>>>>>> Only on the engine VM. >>>>>>>>> Regarding stability - it passes CI so relatively stable, beyond that >>>>>>>>> there are >>>>>>>>> no guarantees. >>>>>>>>> What's the specific problem you're facing with update? Can you >>>>>>>>> elaborate? >>>>>>>>>> Regards, >>>>>>>>>> Tibor >>>>>>>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor < [ >>>>>>>>>> mailto:[email protected] | >>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>> Hi, >>>>>>>>>>> Could you explain how can I use this patch? >>>>>>>>>>> R, >>>>>>>>>>> Tibor >>>>>>>>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor < [ >>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>> Hi, >>>>>>>>>>>> Sorry for my question, but can you tell me please how can I use >>>>>>>>>>>> this patch? >>>>>>>>>>>> Thanks, >>>>>>>>>>>> Regards, >>>>>>>>>>>> Tibor >>>>>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose < [ >>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [ >>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> Could someone help me please ? I can't finish my upgrade process. >>>>>>>>>>>>> [ https://gerrit.ovirt.org/91164 | https://gerrit.ovirt.org/91164 >>>>>>>>>>>>> ] should fix >>>>>>>>>>>>> the error you're facing. >>>>>>>>>>>>> Can you elaborate why this is affecting the upgrade process? >>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>> R >>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ >>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>> I've attached the vdsm and supervdsm logs. But I don't have >>>>>>>>>>>>>>> engine.log here, >>>>>>>>>>>>>>> because that is on hosted engine vm. Should I send that ? >>>>>>>>>>>>>>> Thank you >>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ >>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>>> There's a bug here. Can you log one attaching this engine.log >>>>>>>>>>>>>>>> and also vdsm.log >>>>>>>>>>>>>>>> & supervdsm.log from n3.itsmart.cloud >>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>> I found this: >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,096+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, >>>>>>>>>>>>>>>>> log id: 347435ae >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>>> 'volume2' of >>>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) >>>>>>>>>>>>>>>>> [7715ceda] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>>>>> log id: 6908121d >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), >>>>>>>>>>>>>>>>> log id: 735c6a5f >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>> log id: 6f9e9f58 >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58 >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>>>>> log id: 2ee46967 >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967 >>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 7550e5c >>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, >>>>>>>>>>>>>>>>> log id: 7550e5c >>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>> log id: 120cc68d >>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>>>>> id: 120cc68d >>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>> log id: 118aa264 >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02 WARN >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>>>>> return: >>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g >>>>>>>>>>>>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>>>>> log id: 28d9e255 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>> log id: 4a7b280e >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>>>>> log id: 18adc534 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534 >>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 3451084f >>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02 INFO >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, >>>>>>>>>>>>>>>>> log id: 3451084f >>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02 ERROR >>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) >>>>>>>>>>>>>>>>> [400fa486] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>>> R >>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ >>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>>>>> This doesn't affect the monitoring of state. >>>>>>>>>>>>>>>>>> Any errors in vdsm.log? >>>>>>>>>>>>>>>>>> Or errors in engine.log of the form "Error while refreshing >>>>>>>>>>>>>>>>>> brick statuses for >>>>>>>>>>>>>>>>>> volume" >>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>> Thank you for your fast reply :) >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,574+02 INFO >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>>>> log id: 39adbbb8 >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,768+02 INFO >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>>>>>>> id: 39adbbb8 >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,788+02 INFO >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,892+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,898+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,905+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,911+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,917+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,924+02 WARN >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,925+02 INFO >>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>>>>>>> return: >>>>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>>>>> This happening continuously. >>>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ >>>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>>>>>>> Could you check the engine.log if there are errors related >>>>>>>>>>>>>>>>>>>> to getting >>>>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetails ? >>>>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>>>>>>> Dear Ovirt Users, >>>>>>>>>>>>>>>>>>>>> I've followed up the self-hosted-engine upgrade >>>>>>>>>>>>>>>>>>>>> documentation, I upgraded my 4.1 >>>>>>>>>>>>>>>>>>>>> system to 4.2.3. >>>>>>>>>>>>>>>>>>>>> I upgaded the first node with yum upgrade, it seems >>>>>>>>>>>>>>>>>>>>> working now fine. But since >>>>>>>>>>>>>>>>>>>>> upgrade, the gluster informations seems to displayed >>>>>>>>>>>>>>>>>>>>> incorrect on the admin >>>>>>>>>>>>>>>>>>>>> panel. The volume yellow, and there are red bricks from >>>>>>>>>>>>>>>>>>>>> that node. >>>>>>>>>>>>>>>>>>>>> I've checked in console, I think my gluster is not >>>>>>>>>>>>>>>>>>>>> degraded: >>>>>>>>>>>>>>>>>>>>> root@n1 ~]# gluster volume list >>>>>>>>>>>>>>>>>>>>> volume1 >>>>>>>>>>>>>>>>>>>>> volume2 >>>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume info >>>>>>>>>>>>>>>>>>>>> Volume Name: volume1 >>>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>>>>> Volume Name: volume2 >>>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>>>>> network.ping-timeout: 10 >>>>>>>>>>>>>>>>>>>>> auth.allow: * >>>>>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume status >>>>>>>>>>>>>>>>>>>>> Status of volume: volume1 >>>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>>>>> Task Status of Volume volume1 >>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>>>>> Status of volume: volume2 >>>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968 >>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>>>>> Task Status of Volume volume2 >>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>>>>> I think ovirt can't read valid informations about gluster. >>>>>>>>>>>>>>>>>>>>> I can't contiune upgrade of other hosts until this >>>>>>>>>>>>>>>>>>>>> problem exist. >>>>>>>>>>>>>>>>>>>>> Please help me:) >>>>>>>>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | >>>>>>>>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>>>>>>>>>>> To unsubscribe send an email to [ >>>>>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | >>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>>>> [email protected] ] >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>>> [email protected] ] >>>>>>>>>>> oVirt Code of Conduct: [ >>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ | >>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ ] >>>>>>>>>>> List Archives: > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected]
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

