Nothing in the log really. From restarting the service to error all I get is
2013-12-04 16:51:40 AUDIT [cinder.service] SIGTERM received 2013-12-04 16:51:40 AUDIT [cinder.service] Starting cinder-scheduler node (version 2013.1.4) 2013-12-04 16:51:40 INFO [cinder.openstack.common.rpc.common] Connected to AMQP server on control01:5672 2013-12-04 16:51:40 INFO [cinder.openstack.common.rpc.common] Connected to AMQP server on control01:5672 2013-12-04 16:52:30 WARNING [cinder.scheduler.host_manager] service is down or disabled. 2013-12-04 16:52:30 ERROR [cinder.scheduler.manager] Failed to schedule_create_volume: No valid host was found. That is all I have, no stack trace or anything. Same exact thing on all three controllers. Byron From: John Griffith [mailto:john.griff...@solidfire.com] Sent: Wednesday, December 04, 2013 4:41 PM To: Byron Briggs Cc: openstack@lists.openstack.org; SYSADMIN Subject: Re: [Openstack] [CINDER]Cinder Volume Creation On Wed, Dec 4, 2013 at 2:26 PM, Byron Briggs <byron.bri...@ait.com <mailto:byron.bri...@ait.com> > wrote: root@Compute01-CodeNode:/etc/cinder# vgs File descriptor 3 (/usr/share/bash-completion/completions) leaked on vgs invocation. Parent PID 10778: -bash VG #PV #LV #SN Attr VSize VFree cinder-volumes 1 0 0 wz--n- 100.00g 100.00g My control nodes don't but that shouldn't matter from my understanding. root@control01:/etc/cinder# vgs VG #PV #LV #SN Attr VSize VFree control01-vg 1 2 0 wz--n- 931.27g 44.00m From: John Griffith [mailto:john.griff...@solidfire.com <mailto:john.griff...@solidfire.com> ] Sent: Wednesday, December 04, 2013 3:20 PM To: Byron Briggs Cc: openstack@lists.openstack.org <mailto:openstack@lists.openstack.org> ; SYSADMIN Subject: Re: [Openstack] [CINDER]Cinder Volume Creation On Wed, Dec 4, 2013 at 12:43 PM, Byron Briggs <byron.bri...@ait.com <mailto:byron.bri...@ait.com> > wrote: When running cinder create --display_name test 10 >From the compute01NovaCompute.dmz-pod2 Node listed below. It is also running xenapi for a xenserver. ( Control01,2,3 are an HA cluster sitting behind haproxy/keepalivd running all the communication and schedulers -all that is working well.) I get this error in my cinder-volume.log on the control nodes(since they run schedulers) There is nothing else to go off of other then "ERROR" on the volume status. 2013-12-04 12:55:32 WARNING [cinder.scheduler.host_manager] service is down or disabled. 2013-12-04 12:55:32 ERROR [cinder.scheduler.manager] Failed to schedule_create_volume: No valid host was found. In case you can't see excel below. 192.168.220.40 Is the proxy being distributed into the three control nodes. control01.dmz-pod2(192.168.220.41) ->cinder-api,cinder-scheduler control02.dmz-pod2(192.168.220.42) ->cinder-api,cinder-scheduler control03.dmz-pod2(192.168.220.43) ->cinder-api,cinder-scheduler compute01NovaCompute.dmz-pod2(192.168.220.101) ->cinder-volume Server Name cinder-api cinder-scheduler cinder-volume control01.dmz-pod2 YES YES NO control02.dmz-pod2 YES YES NO control03.dmz-pod2 YES YES NO compute01NovaCompute.dmz-pod2 NO NO NO All services are running with no log errors. Here is my configs compute01NovaCompute.dmz-pod2 /etc/cinder/api-paste.ini [filter:authtoken] paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory service_protocol = http service_host = %SERVICE_TENANT_NAME% service_port = 5000 auth_host = 127.0.0.1 auth_port = 35357 auth_protocol = http admin_tenant_name = services admin_user = %SERVICE_USER% admin_password = %SERVICE_PASSWORD% signing_dir = /var/lib/cinder /etc/cinder/cinder.conf [DEFAULT] iscsi_ip_address=192.168.220.101 rabbit_ha_queues=True rabbit_hosts=control01:5672,control02:5672,control03:5672 rabbit_userid=openstack_rabbit_user rabbit_password=openstack_rabbit_password sql_connection = mysql://cinder:cinder_pass@192.168.220.40/cinder <http://cinder:cinder_pass@192.168.220.40/cinder> rootwrap_config = /etc/cinder/rootwrap.conf api_paste_confg = /etc/cinder/api-paste.ini iscsi_helper = tgtadm volume_name_template = volume-%s volume_group = cinder-volumes verbose = True auth_strategy = keystone state_path = /var/lib/cinder lock_path = /var/lock/cinder volumes_dir = /var/lib/cinder/volumes pvscan File descriptor 3 (/usr/share/bash-completion/completions) leaked on pvscan invocation. Parent PID 10778: -bash PV /dev/xvdb VG cinder-volumes lvm2 [100.00 GiB / 100.00 GiB free] Total: 1 [100.00 GiB] / in use: 1 [100.00 GiB] / in no VG: 0 [0 ] control01,2,3 /etc/cinder/api-paste.ini [filter:authtoken] paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory service_protocol = http service_host = 192.168.220.40 service_port = 5000 auth_host = 192.168.220.40 auth_port = 35357 auth_protocol = http admin_tenant_name = services admin_user = cinder admin_password = keystone_admin signing_dir = /var/lib/cinder /etc/cinder/conder.conf [DEFAULT] sql_idle_timeout=30 rabbit_ha_queues=True rabbit_hosts=control01:5672,control02:5672,control03:5672 rabbit_userid=openstack_rabbit_user rabbit_password=openstack_rabbit_password sql_connection = mysql://cinder:cinder_pass@192.168.220.40/cinder <http://cinder:cinder_pass@192.168.220.40/cinder> osapi_volume_listen = 192.168.220.41 rootwrap_config = /etc/cinder/rootwrap.conf api_paste_confg = /etc/cinder/api-paste.ini iscsi_helper = tgtadm volume_name_template = volume-%s volume_group = nova-volumes verbose = True auth_strategy = keystone state_path = /var/lib/cinder lock_path = /var/lock/cinder volumes_dir = /var/lib/cinder/volumes Grizzly Release Any ideas on where to look more into the issue or something with my config? _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org <mailto:openstack@lists.openstack.org> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack So this happens in a couple of situations, the most common is when there's not enough space/capacity being reported by the configured backend driver to allocate the amount of space you're requesting. Try a "sudo vgs" and verify that you have enough capacity on your backing store (nova-volumes) to deploy a 10G volume. Ooops, sorry I didn't catch the separate controller node and only saw the volume_group setting there. Any chance you could link a paste-bin to the cinder-scheduler logs?
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack