[root@plato ~]# tail /var/log/cinder/volume.log 2018-03-20 22:10:50.258 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:00.259 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:10.260 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:18.826 2572 WARNING cinder.volume.manager [req-b519dd31-2a7a-4188-835a-1a6d3ea9b7b0 - - - - -] Update driver status failed: (config name lvm) is uninitialized. 2018-03-20 22:11:20.262 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:30.263 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:40.264 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:11:50.267 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:12:00.268 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down". 2018-03-20 22:12:10.270 2572 ERROR cinder.service [-] Manager for service cinder-volume plato.spots.onsite@lvm is reporting problems, not sending heartbeat. Service will appear "down".
The nova compute nodes are separate from the controller. Controller: [root@plato ~(keystone_admin)]# openstack service list +----------------------------------+------------+-----------+ | ID | Name | Type | +----------------------------------+------------+-----------+ | 0a6ba1ef756f4a96828bee8f670c31cb | aodh | alarming | | 129820d1c19241978c2d0c3018691af1 | cinderv2 | volumev2 | | 26dfd043bfa041288580a98802751bb6 | placement | placement | | 37c0e999d688402b9ce6705e513da65e | cinder | volume | | 4b6e8f79aeac4a4690a6d5ffb1086886 | ceilometer | metering | | 4fff3f96226d4f16803bfa3fc4502738 | keystone | identity | | 5e050eafaa9046f29f27ef8db9dda05e | neutron | network | | 695363a6e692447b9f498c8831ab7a77 | gnocchi | metric | | 7ff1ddadae5e4b52ad4cb11a94b52b1a | glance | image | | b0a84ca5c85d47d685ba5d60e0e48f93 | nova | compute | | f4323ff470cc4d2ebbfccce6827483c5 | cinderv3 | volumev3 | +----------------------------------+------------+—————+ And separate compute nodes. I am very grateful for your continued help! FV > On Mar 20, 2018, at 10:10 PM, Remo Mattei <r...@italy1.com> wrote: > > Ahh Packstack does not have pcs.. that’s only OOO. > And what does your /var/log/cinder/cinder-volume say? > > Is this all in one? Split? Looks like your cinder is having some issues.. > check out targetcli here is a ref page > > https://www.certdepot.net/rhel7-configure-iscsi-target-initiator-persistently/ > > https://docs.openstack.org/mitaka/install-guide-rdo/cinder-storage-install.html > > > >> On Mar 20, 2018, at 10:03 PM, Father Vlasie <fv@spots.school> wrote: >> >> RDO PackStack >> >> https://www.rdoproject.org/install/packstack/ >> >> >>> On Mar 20, 2018, at 9:35 PM, r...@italy1.com wrote: >>> >>> How did you install OpenStack? >>> >>> dal mio iPhone X >>> >>>> Il giorno 20 mar 2018, alle ore 18:29, Father Vlasie <fv@spots.school> ha >>>> scritto: >>>> >>>> [root@plato ~]# pcs status >>>> -bash: pcs: command not found >>>> >>>> >>>>> On Mar 20, 2018, at 6:28 PM, Remo Mattei <r...@italy1.com> wrote: >>>>> >>>>> Looks like your pacemaker is not running check that out! >>>>> >>>>> sudo pcs status >>>>> >>>>>> On Mar 20, 2018, at 6:24 PM, Father Vlasie <fv@spots.school> wrote: >>>>>> >>>>>> Your help is much appreciated! Thank you. >>>>>> >>>>>> The cinder service is running on the controller node and it is using a >>>>>> disk partition not the loopback device, I did change the default >>>>>> configuration during install with PackStack. >>>>>> >>>>>> [root@plato ~]# pvs >>>>>> PV VG Fmt Attr PSize PFree >>>>>> /dev/vda3 centos lvm2 a-- 1022.80g 4.00m >>>>>> /dev/vdb1 cinder-volumes lvm2 a-- <10.00t <511.85g >>>>>> >>>>>> [root@plato ~]# lvchange -a y volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >>>>>> Volume group "volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5" not found >>>>>> Cannot process volume group volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >>>>>> >>>>>> [root@plato ~]# lvchange -a y cinder-volumes >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) >>>>>> transaction_id is 0, while expected 72. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> On Mar 20, 2018, at 6:05 PM, Vagner Farias <vfar...@redhat.com> wrote: >>>>>>> >>>>>>> Will "lvchange -a y lvname" activate it? >>>>>>> >>>>>>> If not, considering that you're using Pike on Centos, there's a chance >>>>>>> you may be using the cinder-volumes backed by a loopback file. I guess >>>>>>> both packstack & tripleo will configure this by default if you don't >>>>>>> change the configuration. At least tripleo won't configure this >>>>>>> loopback device to be activated automatically on boot. An option would >>>>>>> be to include lines like the following in /etc/rc.d/rc.local: >>>>>>> >>>>>>> losetup /dev/loop0 /var/lib/cinder/cinder-volumes >>>>>>> vgscan >>>>>>> >>>>>>> Last but not least, if this is actually the case, I wouldn't recommend >>>>>>> using loopback devices for LVM SCSI driver. In fact, if you can use any >>>>>>> other driver capable of delivering HA, it'd be better (unless this is >>>>>>> some POC or an environment without tight SLAs). >>>>>>> >>>>>>> Vagner Farias >>>>>>> >>>>>>> >>>>>>> Em ter, 20 de mar de 2018 21:24, Father Vlasie <fv@spots.school> >>>>>>> escreveu: >>>>>>> Here is the output of lvdisplay: >>>>>>> >>>>>>> [root@plato ~]# lvdisplay >>>>>>> --- Logical volume --- >>>>>>> LV Name cinder-volumes-pool >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato, 2018-02-01 13:33:51 -0800 >>>>>>> LV Pool metadata cinder-volumes-pool_tmeta >>>>>>> LV Pool data cinder-volumes-pool_tdata >>>>>>> LV Status NOT available >>>>>>> LV Size 9.50 TiB >>>>>>> Current LE 2490368 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e >>>>>>> LV Name volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0 >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato, 2018-02-02 10:18:41 -0800 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 1.00 GiB >>>>>>> Current LE 256 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3 >>>>>>> LV Name volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3 >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4 >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato, 2018-02-02 10:26:46 -0800 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 1.00 GiB >>>>>>> Current LE 256 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693 >>>>>>> LV Name volume-ee107488-2559-4116-aa7b-0da02fd5f693 >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 40.00 GiB >>>>>>> Current LE 10240 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734 >>>>>>> LV Name volume-d6f0260d-21b5-43e7-afe5-84e0502fa734 >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 40.00 GiB >>>>>>> Current LE 10240 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147 >>>>>>> LV Name volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147 >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 4.00 GiB >>>>>>> Current LE 1024 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path >>>>>>> /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >>>>>>> LV Name volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >>>>>>> VG Name cinder-volumes >>>>>>> LV UUID IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700 >>>>>>> LV Pool name cinder-volumes-pool >>>>>>> LV Status NOT available >>>>>>> LV Size 40.00 GiB >>>>>>> Current LE 10240 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path /dev/centos/root >>>>>>> LV Name root >>>>>>> VG Name centos >>>>>>> LV UUID nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800 >>>>>>> LV Status available >>>>>>> # open 1 >>>>>>> LV Size 50.00 GiB >>>>>>> Current LE 12800 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> - currently set to 8192 >>>>>>> Block device 253:0 >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path /dev/centos/swap >>>>>>> LV Name swap >>>>>>> VG Name centos >>>>>>> LV UUID Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800 >>>>>>> LV Status available >>>>>>> # open 2 >>>>>>> LV Size 8.12 GiB >>>>>>> Current LE 2080 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> - currently set to 8192 >>>>>>> Block device 253:1 >>>>>>> >>>>>>> --- Logical volume --- >>>>>>> LV Path /dev/centos/home >>>>>>> LV Name home >>>>>>> VG Name centos >>>>>>> LV UUID lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg >>>>>>> LV Write Access read/write >>>>>>> LV Creation host, time localhost, 2018-01-22 09:50:39 -0800 >>>>>>> LV Status available >>>>>>> # open 1 >>>>>>> LV Size 964.67 GiB >>>>>>> Current LE 246956 >>>>>>> Segments 1 >>>>>>> Allocation inherit >>>>>>> Read ahead sectors auto >>>>>>> - currently set to 8192 >>>>>>> Block device 253:2 >>>>>>> >>>>>>> >>>>>>>> On Mar 20, 2018, at 4:51 PM, Remo Mattei <r...@italy1.com> wrote: >>>>>>>> >>>>>>>> I think you need to provide a bit of additional info. Did you look at >>>>>>>> the logs? What version of os are you running? Etc. >>>>>>>> >>>>>>>> Inviato da iPhone >>>>>>>> >>>>>>>>> Il giorno 20 mar 2018, alle ore 16:15, Father Vlasie >>>>>>>>> <fv@spots.school> ha scritto: >>>>>>>>> >>>>>>>>> Hello everyone, >>>>>>>>> >>>>>>>>> I am in need of help with my Cinder volumes which have all become >>>>>>>>> unavailable. >>>>>>>>> >>>>>>>>> Is there anyone who would be willing to log in to my system and have >>>>>>>>> a look? >>>>>>>>> >>>>>>>>> My cinder volumes are listed as "NOT available" and my attempts to >>>>>>>>> mount them have been in vain. I have tried: vgchange -a y >>>>>>>>> >>>>>>>>> with result showing as: 0 logical volume(s) in volume group >>>>>>>>> "cinder-volumes" now active >>>>>>>>> >>>>>>>>> I am a bit desperate because some of the data is critical and, I am >>>>>>>>> ashamed to say, I do not have a backup. >>>>>>>>> >>>>>>>>> Any help or suggestions would be very much appreciated. >>>>>>>>> >>>>>>>>> FV >>>>>>>>> _______________________________________________ >>>>>>>>> Mailing list: >>>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>>>>> Post to : openstack@lists.openstack.org >>>>>>>>> Unsubscribe : >>>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Mailing list: >>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>>> Post to : openstack@lists.openstack.org >>>>>>> Unsubscribe : >>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>> >>>>> >>>> >> > _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack