[root@plato ~]# tail  /var/log/cinder/volume.log 
2018-03-20 22:10:50.258 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:00.259 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:10.260 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:18.826 2572 WARNING cinder.volume.manager 
[req-b519dd31-2a7a-4188-835a-1a6d3ea9b7b0 - - - - -] Update driver status 
failed: (config name lvm) is uninitialized.
2018-03-20 22:11:20.262 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:30.263 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:40.264 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:11:50.267 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:12:00.268 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".
2018-03-20 22:12:10.270 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear "down".

The nova compute nodes are separate from the controller.

Controller:

[root@plato ~(keystone_admin)]# openstack service list
+----------------------------------+------------+-----------+
| ID                               | Name       | Type      |
+----------------------------------+------------+-----------+
| 0a6ba1ef756f4a96828bee8f670c31cb | aodh       | alarming  |
| 129820d1c19241978c2d0c3018691af1 | cinderv2   | volumev2  |
| 26dfd043bfa041288580a98802751bb6 | placement  | placement |
| 37c0e999d688402b9ce6705e513da65e | cinder     | volume    |
| 4b6e8f79aeac4a4690a6d5ffb1086886 | ceilometer | metering  |
| 4fff3f96226d4f16803bfa3fc4502738 | keystone   | identity  |
| 5e050eafaa9046f29f27ef8db9dda05e | neutron    | network   |
| 695363a6e692447b9f498c8831ab7a77 | gnocchi    | metric    |
| 7ff1ddadae5e4b52ad4cb11a94b52b1a | glance     | image     |
| b0a84ca5c85d47d685ba5d60e0e48f93 | nova       | compute   |
| f4323ff470cc4d2ebbfccce6827483c5 | cinderv3   | volumev3  |
+----------------------------------+------------+—————+

And separate compute nodes.

I am very grateful for your continued help!

FV

> On Mar 20, 2018, at 10:10 PM, Remo Mattei <r...@italy1.com> wrote:
> 
> Ahh Packstack does not have pcs.. that’s only OOO. 
> And what does your /var/log/cinder/cinder-volume say?
>  
> Is this all in one? Split? Looks like your cinder is having some issues..  
> check out targetcli here is a ref page
> 
> https://www.certdepot.net/rhel7-configure-iscsi-target-initiator-persistently/
> 
> https://docs.openstack.org/mitaka/install-guide-rdo/cinder-storage-install.html
> 
> 
> 
>> On Mar 20, 2018, at 10:03 PM, Father Vlasie <fv@spots.school> wrote:
>> 
>> RDO PackStack
>> 
>> https://www.rdoproject.org/install/packstack/
>> 
>> 
>>> On Mar 20, 2018, at 9:35 PM, r...@italy1.com wrote:
>>> 
>>> How did you install OpenStack? 
>>> 
>>>  dal mio iPhone X 
>>> 
>>>> Il giorno 20 mar 2018, alle ore 18:29, Father Vlasie <fv@spots.school> ha 
>>>> scritto:
>>>> 
>>>> [root@plato ~]# pcs status
>>>> -bash: pcs: command not found
>>>> 
>>>> 
>>>>> On Mar 20, 2018, at 6:28 PM, Remo Mattei <r...@italy1.com> wrote:
>>>>> 
>>>>> Looks like your pacemaker is not running check that out! 
>>>>> 
>>>>> sudo pcs status 
>>>>> 
>>>>>> On Mar 20, 2018, at 6:24 PM, Father Vlasie <fv@spots.school> wrote:
>>>>>> 
>>>>>> Your help is much appreciated! Thank you.
>>>>>> 
>>>>>> The cinder service is running on the controller node and it is using a 
>>>>>> disk partition not the loopback device, I did change the default 
>>>>>> configuration during install with PackStack.
>>>>>> 
>>>>>> [root@plato ~]# pvs
>>>>>> PV         VG             Fmt  Attr PSize    PFree   
>>>>>> /dev/vda3  centos         lvm2 a--  1022.80g    4.00m
>>>>>> /dev/vdb1  cinder-volumes lvm2 a--   <10.00t <511.85g
>>>>>> 
>>>>>> [root@plato ~]# lvchange -a y volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>> Volume group "volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5" not found
>>>>>> Cannot process volume group volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>> 
>>>>>> [root@plato ~]# lvchange -a y cinder-volumes
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>> transaction_id is 0, while expected 72.
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> On Mar 20, 2018, at 6:05 PM, Vagner Farias <vfar...@redhat.com> wrote:
>>>>>>> 
>>>>>>> Will "lvchange -a y lvname" activate it?
>>>>>>> 
>>>>>>> If not, considering that you're using Pike on Centos, there's a chance 
>>>>>>> you may be using the cinder-volumes backed by a loopback file.  I guess 
>>>>>>> both packstack & tripleo will configure this by default if you don't 
>>>>>>> change the configuration. At least tripleo won't configure this 
>>>>>>> loopback device to be activated automatically on boot. An option would 
>>>>>>> be to include lines like the following in /etc/rc.d/rc.local:
>>>>>>> 
>>>>>>> losetup /dev/loop0 /var/lib/cinder/cinder-volumes
>>>>>>> vgscan
>>>>>>> 
>>>>>>> Last but not least, if this is actually the case, I wouldn't recommend 
>>>>>>> using loopback devices for LVM SCSI driver. In fact, if you can use any 
>>>>>>> other driver capable of delivering HA, it'd be better (unless this is 
>>>>>>> some POC or an environment without tight SLAs). 
>>>>>>> 
>>>>>>> Vagner Farias
>>>>>>> 
>>>>>>> 
>>>>>>> Em ter, 20 de mar de 2018 21:24, Father Vlasie <fv@spots.school> 
>>>>>>> escreveu:
>>>>>>> Here is the output of lvdisplay:
>>>>>>> 
>>>>>>> [root@plato ~]# lvdisplay
>>>>>>> --- Logical volume ---
>>>>>>> LV Name                cinder-volumes-pool
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato, 2018-02-01 13:33:51 -0800
>>>>>>> LV Pool metadata       cinder-volumes-pool_tmeta
>>>>>>> LV Pool data           cinder-volumes-pool_tdata
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                9.50 TiB
>>>>>>> Current LE             2490368
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>>>>>> LV Name                volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato, 2018-02-02 10:18:41 -0800
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                1.00 GiB
>>>>>>> Current LE             256
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>>>>>> LV Name                volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato, 2018-02-02 10:26:46 -0800
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                1.00 GiB
>>>>>>> Current LE             256
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>>>>>> LV Name                volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                40.00 GiB
>>>>>>> Current LE             10240
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>>>>>> LV Name                volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                40.00 GiB
>>>>>>> Current LE             10240
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>>>>>> LV Name                volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                4.00 GiB
>>>>>>> Current LE             1024
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                
>>>>>>> /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>> LV Name                volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>> VG Name                cinder-volumes
>>>>>>> LV UUID                IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700
>>>>>>> LV Pool name           cinder-volumes-pool
>>>>>>> LV Status              NOT available
>>>>>>> LV Size                40.00 GiB
>>>>>>> Current LE             10240
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                /dev/centos/root
>>>>>>> LV Name                root
>>>>>>> VG Name                centos
>>>>>>> LV UUID                nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>>>>>> LV Status              available
>>>>>>> # open                 1
>>>>>>> LV Size                50.00 GiB
>>>>>>> Current LE             12800
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> - currently set to     8192
>>>>>>> Block device           253:0
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                /dev/centos/swap
>>>>>>> LV Name                swap
>>>>>>> VG Name                centos
>>>>>>> LV UUID                Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>>>>>> LV Status              available
>>>>>>> # open                 2
>>>>>>> LV Size                8.12 GiB
>>>>>>> Current LE             2080
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> - currently set to     8192
>>>>>>> Block device           253:1
>>>>>>> 
>>>>>>> --- Logical volume ---
>>>>>>> LV Path                /dev/centos/home
>>>>>>> LV Name                home
>>>>>>> VG Name                centos
>>>>>>> LV UUID                lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg
>>>>>>> LV Write Access        read/write
>>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:39 -0800
>>>>>>> LV Status              available
>>>>>>> # open                 1
>>>>>>> LV Size                964.67 GiB
>>>>>>> Current LE             246956
>>>>>>> Segments               1
>>>>>>> Allocation             inherit
>>>>>>> Read ahead sectors     auto
>>>>>>> - currently set to     8192
>>>>>>> Block device           253:2
>>>>>>> 
>>>>>>> 
>>>>>>>> On Mar 20, 2018, at 4:51 PM, Remo Mattei <r...@italy1.com> wrote:
>>>>>>>> 
>>>>>>>> I think you need to provide a bit of additional info. Did you look at 
>>>>>>>> the logs? What version of os are you running? Etc.
>>>>>>>> 
>>>>>>>> Inviato da iPhone
>>>>>>>> 
>>>>>>>>> Il giorno 20 mar 2018, alle ore 16:15, Father Vlasie 
>>>>>>>>> <fv@spots.school> ha scritto:
>>>>>>>>> 
>>>>>>>>> Hello everyone,
>>>>>>>>> 
>>>>>>>>> I am in need of help with my Cinder volumes which have all become 
>>>>>>>>> unavailable.
>>>>>>>>> 
>>>>>>>>> Is there anyone who would be willing to log in to my system and have 
>>>>>>>>> a look?
>>>>>>>>> 
>>>>>>>>> My cinder volumes are listed as "NOT available" and my attempts to 
>>>>>>>>> mount them have been in vain. I have tried: vgchange -a y
>>>>>>>>> 
>>>>>>>>> with result showing as:  0 logical volume(s) in volume group 
>>>>>>>>> "cinder-volumes" now active
>>>>>>>>> 
>>>>>>>>> I am a bit desperate because some of the data is critical and, I am 
>>>>>>>>> ashamed to say, I do not have a backup.
>>>>>>>>> 
>>>>>>>>> Any help or suggestions would be very much appreciated.
>>>>>>>>> 
>>>>>>>>> FV
>>>>>>>>> _______________________________________________
>>>>>>>>> Mailing list: 
>>>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>>>> Post to     : openstack@lists.openstack.org
>>>>>>>>> Unsubscribe : 
>>>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> Mailing list: 
>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> Post to     : openstack@lists.openstack.org
>>>>>>> Unsubscribe : 
>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>> 
>>>>> 
>>>> 
>> 
> 


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to