Dear Sahina, 

Is there any changes with this bug? 

Still I haven't finish my upgrade process that i've started on 9th may:( 

Please help me if you can. 

Thanks 

Tibor 

----- 2018. máj.. 18., 9:29, Demeter Tibor <[email protected]> írta: 

> Hi,

> I have to update the engine again?

> Thanks,

> R
> Tibor

> ----- 2018. máj.. 18., 6:47, Sahina Bose <[email protected]> írta:

>> Thanks for reporting this. [ https://gerrit.ovirt.org/91375 |
>> https://gerrit.ovirt.org/91375 ] fixes this. I've re-opened bug [
>> https://bugzilla.redhat.com/show_bug.cgi?id=1574508 |
>> https://bugzilla.redhat.com/show_bug.cgi?id=1574508 ]

>> On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor < [ 
>> mailto:[email protected] |
>> [email protected] ] > wrote:

>>> Hi,

>>> 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos

>>> Firstly, I did a yum update "ovirt-*-setup*"
>>> second, I have ran engine-setup to upgrade.

>>> I didn't remove the old repos, just installed the nightly repo.

>>> Thank you again,

>>> Regards,

>>> Tibor

>>> ----- 2018. máj.. 17., 15:02, Sahina Bose < [ mailto:[email protected] |
>>> [email protected] ] > írta:

>>>> It doesn't look like the patch was applied. Still see the same error in
>>>> engine.log
>>>> "Error while refreshing brick statuses for volume 'volume1' of cluster 
>>>> 'C6220':
>>>> null"\

>>>> Did you use engine-setup to upgrade? What's the version of ovirt-engine
>>>> currently installed?

>>>> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [ 
>>>> mailto:[email protected] |
>>>> [email protected] ] > wrote:

>>>>> Hi,

>>>>> sure,

>>>>> Thank you for your time!

>>>>> R
>>>>> Tibor

>>>>> ----- 2018. máj.. 17., 12:19, Sahina Bose < [ mailto:[email protected] |
>>>>> [email protected] ] > írta:

>>>>>> [+users]

>>>>>> Can you provide the engine.log to see why the monitoring is not working 
>>>>>> here.
>>>>>> thanks!

>>>>>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [ 
>>>>>> mailto:[email protected] |
>>>>>> [email protected] ] > wrote:

>>>>>>> Hi,

>>>>>>> Meanwhile, I did the upgrade engine, but the gluster state is same on 
>>>>>>> my first
>>>>>>> node.
>>>>>>> I've attached some screenshot of my problem.

>>>>>>> Thanks

>>>>>>> Tibor

>>>>>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor < [ 
>>>>>>> mailto:[email protected] |
>>>>>>> [email protected] ] > írta Hi,

>>>>>>>> If 4.3.4 will release, i just have to remove the nightly repo and 
>>>>>>>> update to
>>>>>>>> stable?

>>>>>>>> I'm sorry for my terrible English, I try to explain what was my 
>>>>>>>> problem with
>>>>>>>> update.
>>>>>>>> I'm upgraded from 4.1.8.

>>>>>>>> I followed up the official hosted-engine update documentation, that 
>>>>>>>> was not
>>>>>>>> clear me, because it has referenced to a lot of old thing (i think).
>>>>>>>> [ https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ |
>>>>>>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ ]
>>>>>>>> [
>>>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>>>>> |
>>>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>>>>> ]

>>>>>>>> Maybe it need to update, because I had a lot of question under upgrade 
>>>>>>>> and I was
>>>>>>>> not sure in all of necessary steps. For example, If I need to 
>>>>>>>> installing the
>>>>>>>> new, 4.2 repo on the hosts, then need to remove the old repo from that?
>>>>>>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an 
>>>>>>>> "Updatehost"
>>>>>>>> menu in the GUI? So, maybe it outdated.
>>>>>>>> Since upgrade hosted engine, and the first node, I have problems with 
>>>>>>>> gluster.
>>>>>>>> It seems to working fine if you check it from console "gluster volume 
>>>>>>>> status,
>>>>>>>> etc" but not on the Gui, because now it yellow, and the brick reds in 
>>>>>>>> the first
>>>>>>>> node.

>>>>>>>> Previously I did a mistake with glusterfs, my gluster config was 
>>>>>>>> wrong. I have
>>>>>>>> corrected them, but it did not helped to me,gluster bricks are reds on 
>>>>>>>> my first
>>>>>>>> node yet....

>>>>>>>> Now I try to upgrade to nightly, but I'm affraid, because it a living,
>>>>>>>> productive system, and I don't have downtime. I hope it will help me.

>>>>>>>> Thanks for all,

>>>>>>>> Regards,
>>>>>>>> Tibor Demeter

>>>>>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose < [ mailto:[email protected] |
>>>>>>>> [email protected] ] > írta:

>>>>>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor < [ 
>>>>>>>>> mailto:[email protected] |
>>>>>>>>> [email protected] ] > wrote:

>>>>>>>>>> Hi,

>>>>>>>>>> is it a different, unstable repo? I have a productive cluster, how 
>>>>>>>>>> is safe that?
>>>>>>>>>> I don't have any experience with nightly build. How can I use this? 
>>>>>>>>>> It have to
>>>>>>>>>> install to the engine VM or all of my hosts?
>>>>>>>>>> Thanks in advance for help me..

>>>>>>>>> Only on the engine VM.

>>>>>>>>> Regarding stability - it passes CI so relatively stable, beyond that 
>>>>>>>>> there are
>>>>>>>>> no guarantees.

>>>>>>>>> What's the specific problem you're facing with update? Can you 
>>>>>>>>> elaborate?

>>>>>>>>>> Regards,

>>>>>>>>>> Tibor

>>>>>>>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor < [ 
>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>> Hi,

>>>>>>>>>>> Could you explain how can I use this patch?

>>>>>>>>>>> R,
>>>>>>>>>>> Tibor

>>>>>>>>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor < [ 
>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>> Hi,

>>>>>>>>>>>> Sorry for my question, but can you tell me please how can I use 
>>>>>>>>>>>> this patch?

>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> Tibor
>>>>>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose < [ 
>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [ 
>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>> [email protected] ] > wrote:

>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>> Could someone help me please ? I can't finish my upgrade process.

>>>>>>>>>>>>> [ https://gerrit.ovirt.org/91164 | https://gerrit.ovirt.org/91164 
>>>>>>>>>>>>> ] should fix
>>>>>>>>>>>>> the error you're facing.

>>>>>>>>>>>>> Can you elaborate why this is affecting the upgrade process?

>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>> R
>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ 
>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>>> I've attached the vdsm and supervdsm logs. But I don't have 
>>>>>>>>>>>>>>> engine.log here,
>>>>>>>>>>>>>>> because that is on hosted engine vm. Should I send that ?

>>>>>>>>>>>>>>> Thank you

>>>>>>>>>>>>>>> Regards,

>>>>>>>>>>>>>>> Tibor
>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ 
>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>>>>>> There's a bug here. Can you log one attaching this engine.log 
>>>>>>>>>>>>>>>> and also vdsm.log
>>>>>>>>>>>>>>>> & supervdsm.log from n3.itsmart.cloud

>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ 
>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>> [email protected] ] > wrote:

>>>>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>>>>> I found this:

>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,096+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
>>>>>>>>>>>>>>>>> log id: 347435ae
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
>>>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>>>>> 'volume2' of
>>>>>>>>>>>>>>>>> cluster 'C6220': null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
>>>>>>>>>>>>>>>>> [7715ceda] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>>>>> log id: 6908121d
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
>>>>>>>>>>>>>>>>> log id: 735c6a5f
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 6f9e9f58
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>>>>> log id: 2ee46967
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 7550e5c
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
>>>>>>>>>>>>>>>>> log id: 7550e5c
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
>>>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>>>>> 'volume1' of
>>>>>>>>>>>>>>>>> cluster 'C6220': null
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 120cc68d
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand,
>>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 
>>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 
>>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>>>>> id: 120cc68d
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 118aa264
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02 WARN
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
>>>>>>>>>>>>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264

>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>>>>> log id: 28d9e255
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 4a7b280e
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>>>>> log id: 18adc534
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 
>>>>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 3451084f
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02 INFO
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
>>>>>>>>>>>>>>>>> log id: 3451084f
>>>>>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02 ERROR
>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
>>>>>>>>>>>>>>>>> [400fa486] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>>>>> 'volume1' of
>>>>>>>>>>>>>>>>> cluster 'C6220': null

>>>>>>>>>>>>>>>>> R
>>>>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ 
>>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>>>>>>>> This doesn't affect the monitoring of state.
>>>>>>>>>>>>>>>>>> Any errors in vdsm.log?
>>>>>>>>>>>>>>>>>> Or errors in engine.log of the form "Error while refreshing 
>>>>>>>>>>>>>>>>>> brick statuses for
>>>>>>>>>>>>>>>>>> volume"

>>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ 
>>>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>>>> [email protected] ] > wrote:

>>>>>>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>>>>>>> Thank you for your fast reply :)

>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,574+02 INFO
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>>>> log id: 39adbbb8
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,768+02 INFO
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand,
>>>>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 
>>>>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 
>>>>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>>>>>>> id: 39adbbb8
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,788+02 INFO
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>>>> log id: 738a7261
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,892+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,898+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,905+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,911+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,917+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,924+02 WARN
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate 
>>>>>>>>>>>>>>>>>>> brick
>>>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network 
>>>>>>>>>>>>>>>>>>> as no gluster
>>>>>>>>>>>>>>>>>>> network found in cluster 
>>>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,925+02 INFO
>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>>>>>>>>>>>>>>> log id: 738a7261

>>>>>>>>>>>>>>>>>>> This happening continuously.

>>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ 
>>>>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>>>>> [email protected] ] > írta:

>>>>>>>>>>>>>>>>>>>> Could you check the engine.log if there are errors related 
>>>>>>>>>>>>>>>>>>>> to getting
>>>>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetails ?

>>>>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ 
>>>>>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>>>>>> [email protected] ] > wrote:

>>>>>>>>>>>>>>>>>>>>> Dear Ovirt Users,
>>>>>>>>>>>>>>>>>>>>> I've followed up the self-hosted-engine upgrade 
>>>>>>>>>>>>>>>>>>>>> documentation, I upgraded my 4.1
>>>>>>>>>>>>>>>>>>>>> system to 4.2.3.
>>>>>>>>>>>>>>>>>>>>> I upgaded the first node with yum upgrade, it seems 
>>>>>>>>>>>>>>>>>>>>> working now fine. But since
>>>>>>>>>>>>>>>>>>>>> upgrade, the gluster informations seems to displayed 
>>>>>>>>>>>>>>>>>>>>> incorrect on the admin
>>>>>>>>>>>>>>>>>>>>> panel. The volume yellow, and there are red bricks from 
>>>>>>>>>>>>>>>>>>>>> that node.
>>>>>>>>>>>>>>>>>>>>> I've checked in console, I think my gluster is not 
>>>>>>>>>>>>>>>>>>>>> degraded:

>>>>>>>>>>>>>>>>>>>>> root@n1 ~]# gluster volume list
>>>>>>>>>>>>>>>>>>>>> volume1
>>>>>>>>>>>>>>>>>>>>> volume2
>>>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume info
>>>>>>>>>>>>>>>>>>>>> Volume Name: volume1
>>>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>>>>>>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>>>>>>>>>>>>>>> Status: Started
>>>>>>>>>>>>>>>>>>>>> Snapshot Count: 0
>>>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Options Reconfigured:
>>>>>>>>>>>>>>>>>>>>> transport.address-family: inet
>>>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>>>>> nfs.disable: on
>>>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36
>>>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36
>>>>>>>>>>>>>>>>>>>>> performance.quick-read: off
>>>>>>>>>>>>>>>>>>>>> performance.read-ahead: off
>>>>>>>>>>>>>>>>>>>>> performance.io-cache: off
>>>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>>>>> network.remote-dio: enable
>>>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>>>>> features.shard: on
>>>>>>>>>>>>>>>>>>>>> user.cifs: off
>>>>>>>>>>>>>>>>>>>>> server.allow-insecure: on
>>>>>>>>>>>>>>>>>>>>> Volume Name: volume2
>>>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>>>>>>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>>>>>>>>>>>>>>> Status: Started
>>>>>>>>>>>>>>>>>>>>> Snapshot Count: 0
>>>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>>>> Options Reconfigured:
>>>>>>>>>>>>>>>>>>>>> nfs.disable: on
>>>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>>>>> transport.address-family: inet
>>>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>>>>> network.ping-timeout: 10
>>>>>>>>>>>>>>>>>>>>> auth.allow: *
>>>>>>>>>>>>>>>>>>>>> performance.quick-read: off
>>>>>>>>>>>>>>>>>>>>> performance.read-ahead: off
>>>>>>>>>>>>>>>>>>>>> performance.io-cache: off
>>>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>>>>> network.remote-dio: enable
>>>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>>>>> features.shard: on
>>>>>>>>>>>>>>>>>>>>> user.cifs: off
>>>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36
>>>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36
>>>>>>>>>>>>>>>>>>>>> server.allow-insecure: on
>>>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume status
>>>>>>>>>>>>>>>>>>>>> Status of volume: volume1
>>>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>>>>> Task Status of Volume volume1
>>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>>>>>>>>>>> Status of volume: volume2
>>>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>>>>> Task Status of Volume volume2
>>>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>>>>>>>>>>> I think ovirt can't read valid informations about gluster.
>>>>>>>>>>>>>>>>>>>>> I can't contiune upgrade of other hosts until this 
>>>>>>>>>>>>>>>>>>>>> problem exist.

>>>>>>>>>>>>>>>>>>>>> Please help me:)

>>>>>>>>>>>>>>>>>>>>> Thanks

>>>>>>>>>>>>>>>>>>>>> Regards,

>>>>>>>>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | 
>>>>>>>>>>>>>>>>>>>>> [email protected] ]
>>>>>>>>>>>>>>>>>>>>> To unsubscribe send an email to [ 
>>>>>>>>>>>>>>>>>>>>> mailto:[email protected] |
>>>>>>>>>>>>>>>>>>>>> [email protected] ]

>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | 
>>>>>>>>>>>>>>> [email protected] ]
>>>>>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] |
>>>>>>>>>>>>>>> [email protected] ]

>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] |
>>>>>>>>>>>> [email protected] ]

>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] |
>>>>>>>>>>> [email protected] ]
>>>>>>>>>>> oVirt Code of Conduct: [
>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ |
>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ ]
>>>>>>>>>>> List Archives:

> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to