list other three pool information as below:
root@ubuntu:~# virsh pool-info ac2949d9-16fb-4c54-a551-1cdbeb501f05
Name:           ac2949d9-16fb-4c54-a551-1cdbeb501f05
UUID:           ac2949d9-16fb-4c54-a551-1cdbeb501f05
State:          running
Persistent:     yes
Autostart:      no
Capacity:       811.17 GiB
Allocation:     13.11 GiB
Available:      798.05 GiB

root@ubuntu:~# virsh pool-info d9474b9d-afa1-3737-a13b-df333dae295f
Name:           d9474b9d-afa1-3737-a13b-df333dae295f
UUID:           d9474b9d-afa1-3737-a13b-df333dae295f
State:          running
Persistent:     yes
Autostart:      no
Capacity:       57.67 GiB
Allocation:     9.86 GiB
Available:      47.81 GiB

root@ubuntu:~# virsh pool-info f81769c1-a31b-4202-a3aa-bdf616eef019
Name:           f81769c1-a31b-4202-a3aa-bdf616eef019
UUID:           f81769c1-a31b-4202-a3aa-bdf616eef019
State:          running
Persistent:     yes
Autostart:      no
Capacity:       811.17 GiB
Allocation:     13.11 GiB
Available:      798.05 GiB


On Mon, Apr 22, 2013 at 9:57 AM, Guangjian Liu <guangj...@gmail.com> wrote:

> Wido,
>
> Please see information below
>
> How big is the Ceph cluster? (ceph -s).
> [root@RDR02S01 ~]# ceph -s
>    health HEALTH_OK
>    monmap e1: 1 mons at {a=10.0.0.41:6789/0}, election epoch 1, quorum 0 a
>    osdmap e35: 2 osds: 2 up, 2 in
>     pgmap v17450: 714 pgs: 714 active+clean; 2522 MB data, 10213 MB used,
> 126 GB / 136 GB avail
>    mdsmap e557: 1/1/1 up {0=a=up:active}
>
> And what does libvirt say?
>
> 1. Log on to a hypervisor
> 2. virsh pool-list
> root@ubuntu:~# virsh pool-list
> Name                 State      Autostart
> -----------------------------------------
> 18559bbf-cd1a-3fcc-a4d2-e95cd4f2d78c active     no
> ac2949d9-16fb-4c54-a551-1cdbeb501f05 active     no
> d9474b9d-afa1-3737-a13b-df333dae295f active     no
> f81769c1-a31b-4202-a3aa-bdf616eef019 active     no
>
> 3. virsh pool-info <uuid of ceph pool>
> root@ubuntu:~# virsh pool-info 18559bbf-cd1a-3fcc-a4d2-e95cd4f2d78c
> Name:           18559bbf-cd1a-3fcc-a4d2-e95cd4f2d78c
> UUID:           18559bbf-cd1a-3fcc-a4d2-e95cd4f2d78c
> State:          running
> Persistent:     yes
> Autostart:      no
> Capacity:       136.60 GiB
> *Allocation:     30723.02 TiB*
> Available:      126.63 GiB
>
>
> On Fri, Apr 19, 2013 at 4:43 PM, Wido den Hollander <w...@widodh.nl>wrote:
>
>> Hi,
>>
>>
>> On 04/19/2013 08:48 AM, Guangjian Liu wrote:
>>
>>> Ceph file system can added as Primary storage by RBD, Thanks.
>>>
>>>
>> Great!
>>
>>
>>  Now I meet the problem in creating VM to Ceph.
>>>
>>> First I create computer offer and storage offer with tag "ceph", and
>>> define Ceph Primary storage with tag "ceph".
>>> Then I create VM use "ceph" computer offer, It display exception in Log.
>>> It seems the storage size out off usage.
>>>
>>>
>> That is indeed odd. How big is the Ceph cluster? (ceph -s).
>>
>> And what does libvirt say?
>>
>> 1. Log on to a hypervisor
>> 2. virsh pool-list
>> 3. virsh pool-info <uuid of ceph pool>
>>
>> I'm curious about what libvirt reports as pool size.
>>
>> Wido
>>
>>  Create VM fail
>>> 2013-04-18 10:23:27,559 DEBUG
>>> [storage.allocator.**AbstractStoragePoolAllocator]
>>> (Job-Executor-1:job-24)
>>> Is storage pool shared? true
>>> 2013-04-18 10:23:27,561 DEBUG [cloud.storage.**StorageManagerImpl]
>>> (Job-Executor-1:job-24) Checking pool 209 for storage, totalSize:
>>> 146673582080, *usedBytes: 33780317214998658*, usedPct:
>>>
>>> 230309.48542985672, disable threshold: 0.85
>>> 2013-04-18 10:23:27,561 DEBUG [cloud.storage.**StorageManagerImpl]
>>> (Job-Executor-1:job-24) Insufficient space on pool: 209 since its usage
>>> percentage: 230309.48542985672 has crossed the
>>> pool.storage.capacity.**disablethreshold: 0.85
>>> 2013-04-18 10:23:27,561 DEBUG
>>> [storage.allocator.**FirstFitStoragePoolAllocator]
>>> (Job-Executor-1:job-24)
>>> FirstFitStoragePoolAllocator returning 0 suitable storage pools
>>> 2013-04-18 10:23:27,561 DEBUG [cloud.deploy.FirstFitPlanner]
>>> (Job-Executor-1:job-24) No suitable pools found for volume:
>>> Vol[8|vm=8|ROOT] under cluster: 1
>>> 2013-04-18 10:23:27,562 DEBUG [cloud.deploy.FirstFitPlanner]
>>> (Job-Executor-1:job-24) No suitable pools found
>>> 2013-04-18 10:23:27,562 DEBUG [cloud.deploy.FirstFitPlanner]
>>> (Job-Executor-1:job-24) No suitable storagePools found under this
>>> Cluster: 1
>>> 2013-04-18 10:23:27,562 DEBUG [cloud.deploy.FirstFitPlanner]
>>> (Job-Executor-1:job-24) Could not find suitable Deployment Destination
>>> for this VM under any clusters, returning.
>>> 2013-04-18 10:23:27,640 DEBUG [cloud.capacity.**CapacityManagerImpl]
>>> (Job-Executor-1:job-24) VM state transitted from :Starting to Stopped
>>> with event: OperationFailedvm's original host id: null new host id: null
>>> host id before state transition: null
>>> 2013-04-18 10:23:27,645 DEBUG [cloud.vm.UserVmManagerImpl]
>>> (Job-Executor-1:job-24) Destroying vm VM[User|ceph-1] as it failed to
>>> create
>>> management-server.7z 2013-04-18 10:02
>>>
>>> I check CS Web GUI and Database, the storage Allocated 30723TB, but
>>> actually the total size is 138GB.
>>> Inline image 1
>>>
>>>
>>> mysql> select * from storage_pool;
>>> +-----+---------+-------------**-------------------------+----**
>>> ---------------+------+-------**---------+--------+-----------**
>>> -+-------------------+--------**--------+--------------+------**
>>> -----+-----------------+------**---------------+---------+----**
>>> ---------+--------+
>>> | id  | name    | uuid                                 | pool_type
>>>    | port | data_center_id | pod_id | cluster_id | available_bytes   |
>>> capacity_bytes | host_address | user_info | path            | created
>>>            | removed | update_time | status |
>>> +-----+---------+-------------**-------------------------+----**
>>> ---------------+------+-------**---------+--------+-----------**
>>> -+-------------------+--------**--------+--------------+------**
>>> -----+-----------------+------**---------------+---------+----**
>>> ---------+--------+
>>> | 200 | primary | d9474b9d-afa1-3737-a13b-**df333dae295f |
>>> NetworkFilesystem | 2049 |              1 |      1 |          1 |
>>> 10301210624 |    61927849984 | 10.0.0.42    | NULL      |
>>> /export/primary | 2013-04-17 06:25:47 | NULL    | NULL        | Up     |
>>> | 209 | ceph    | 18559bbf-cd1a-3fcc-a4d2-**e95cd4f2d78c | RBD
>>>    | 6789 |              1 |      1 |          1 | 33780317214998658 |
>>> 146673582080 | 10.0.0.41    | :         | rbd             | 2013-04-18
>>> 02:02:00 | NULL    | NULL        | Up     |
>>> +-----+---------+-------------**-------------------------+----**
>>> ---------------+------+-------**---------+--------+-----------**
>>> -+-------------------+--------**--------+--------------+------**
>>> -----+-----------------+------**---------------+---------+----**
>>> ---------+--------+
>>> 2 rows in set (0.00 sec)
>>>
>>>
>>>
>>>
>>>
>>> On Thu, Apr 18, 2013 at 3:54 AM, Wido den Hollander <w...@widodh.nl
>>> <mailto:w...@widodh.nl>> wrote:
>>>
>>>     Hi,
>>>
>>>
>>>     On 04/17/2013 03:01 PM, Guangjian Liu wrote:
>>>
>>>         I still meet the same result.
>>>
>>>         In ubuntu 12.04,
>>>         1. I install libvirt-dev as below,
>>>              apt-get install libvirt-dev
>>>         2. rebuild libvirt, see detail build log in attach.
>>>         root@ubuntu:~/install/libvirt-**__0.10.2# ./autogen.sh
>>>
>>>         running CONFIG_SHELL=/bin/bash /bin/bash ./configure --enable-rbd
>>>         --no-create --no-recursion
>>>         configure: WARNING: unrecognized options: --enable-rbd
>>>
>>>
>>>     The correct option is "--with-storage-rbd"
>>>
>>>     But check the output of configure, it should tell you whether RBD
>>>     was enabled or not.
>>>
>>>     Then verify again if you can create a RBD storage pool manually via
>>>     libvirt.
>>>
>>>     Wido
>>>
>>>         .....
>>>         make
>>>         make install
>>>
>>>
>>>
>>>         On Wed, Apr 17, 2013 at 6:25 PM, Wido den Hollander
>>>         <w...@widodh.nl <mailto:w...@widodh.nl>
>>>         <mailto:w...@widodh.nl <mailto:w...@widodh.nl>>> wrote:
>>>
>>>              Hi,
>>>
>>>
>>>              On 04/17/2013 11:37 AM, Guangjian Liu wrote:
>>>
>>>                  Thanks for your mail, you suggest compile libvirt with
>>>         RBD enable.
>>>                  I already build libvirt-0.10.2.tar.gz as document
>>>         
>>> http://ceph.com/docs/master/__**__rbd/libvirt/<http://ceph.com/docs/master/____rbd/libvirt/>
>>>         
>>> <http://ceph.com/docs/master/_**_rbd/libvirt/<http://ceph.com/docs/master/__rbd/libvirt/>
>>> >
>>>
>>>
>>>                  
>>> <http://ceph.com/docs/master/_**_rbd/libvirt/<http://ceph.com/docs/master/__rbd/libvirt/>
>>>         
>>> <http://ceph.com/docs/master/**rbd/libvirt/<http://ceph.com/docs/master/rbd/libvirt/>>>
>>> in my SERVER C(Ubuntu
>>>                  12.04),
>>>                     Shall I build libvirt-0.10.2.tar.gz with RBD enable?
>>>           use
>>>                  ./configure
>>>                  --enable-rbd instead autogen.sh?
>>>
>>>
>>>              Well, you don't have to add --enable-rbd to configure nor
>>>              autogen.sh, but you have to make sure the development
>>>         libraries for
>>>              librbd are installed.
>>>
>>>              On CentOS do this:
>>>
>>>              yum install librbd-devel
>>>
>>>              And retry autogen.sh for libvirt, it should tell you RBD is
>>>         enabled.
>>>
>>>              Wido
>>>
>>>                  cd libvirt
>>>                  ./autogen.sh
>>>                  make
>>>                  sudo make install
>>>
>>>
>>>
>>>                  On Wed, Apr 17, 2013 at 4:37 PM, Wido den Hollander
>>>                  <w...@widodh.nl <mailto:w...@widodh.nl>
>>>         <mailto:w...@widodh.nl <mailto:w...@widodh.nl>>> wrote:
>>>
>>>                      Hi,
>>>
>>>
>>>                      On 04/17/2013 01:44 AM, Guangjian Liu wrote:
>>>
>>>                          Create rbd primary storage fail in CS 4.0.1
>>>                          Anybody can help about it!
>>>
>>>                          Environment:
>>>                          1. Server A: CS 4.0.1 OS: RHEL 6.2 x86-64
>>>                          2. Server B: Ceph 0.56.4  OS: RHEL 6.2 x86-64
>>>                          3. Server C: KVM/Qemu OS: Ubuntu 12.04
>>>                                 compile libvirt and Qemu as document
>>>                          root@ubuntu:/usr/local/lib# virsh version
>>>                          Compiled against library: libvirt 0.10.2
>>>                          Using library: libvirt 0.10.2
>>>                          Using API: QEMU 0.10.2
>>>                          Running hypervisor: QEMU 1.0.0
>>>
>>>
>>>                      Are you sure both libvirt and Qemu are compiled
>>>         with RBD
>>>                      enabled?
>>>
>>>                      On your CentOS system you should make sure
>>>         librbd-dev is
>>>                      installed during
>>>                      compilation of libvirt and Qemu.
>>>
>>>                      The most important part is the RBD storage pool
>>>         support in
>>>                      libvirt, that
>>>                      should be enabled.
>>>
>>>                      In the e-mail you send me directly I saw this:
>>>
>>>                      root@ubuntu:~/scripts# virsh pool-define
>>>         rbd-pool.xml error:
>>>                      Failed to
>>>                      define pool from rbd-pool.xml error: internal error
>>>         missing
>>>                      backend for
>>>                      pool type 8
>>>
>>>                      That suggest RBD storage pool support is not
>>>         enabled in libvirt.
>>>
>>>                      Wido
>>>
>>>
>>>                         Problem:
>>>
>>>                          create primary storage fail with rbd device.
>>>
>>>                          Fail log:
>>>                          2013-04-16 16:27:14,224 DEBUG
>>>                          [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) createPool Params @
>>>         scheme - rbd
>>>                          storageHost -
>>>                          10.0.0.41 hostPath - /cloudstack port - -1
>>>                          2013-04-16 16:27:14,270 DEBUG
>>>                          [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) In createPool Setting
>>>         poolId -
>>>                          218 uuid -
>>>                          5924a2df-d658-3119-8aba-**____**f90307683206
>>>
>>>         zoneId - 4
>>>
>>>                          podId - 4 poolName -
>>>                          ceph
>>>                          2013-04-16 16:27:14,318 DEBUG
>>>                          [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) creating pool ceph on
>>>           host 18
>>>                          2013-04-16 16:27:14,320 DEBUG
>>>         [agent.transport.Request]
>>>                          (catalina-exec-9:null) Seq 18-1625162275:
>>>         Sending  { Cmd
>>>                          , MgmtId:
>>>                          37528005876872, via: 18, Ver: v1, Flags: 100011,
>>>
>>>         [{"CreateStoragePoolCommand":{**____**"add":true,"pool":{"id":**
>>> __218,__"**
>>>
>>>         uuid":"5924a2df-d658-3119-**__**__8aba-f90307683206","host":"**
>>> 10.____**
>>>
>>>         0.0.41","path":"cloudstack","***____*userInfo":":","port":**
>>> 6789,"__**
>>>
>>>         type":"RBD"},"localPath":"/**_**___mnt//3cf4f0e8-781d-39d8-**
>>> b81c-__*__*
>>>
>>>
>>>
>>>                          9896da212335","wait":0}}]
>>>                          }
>>>                          2013-04-16 16:27:14,323 DEBUG
>>>         [agent.transport.Request]
>>>                          (AgentManager-Handler-2:null) Seq 18-1625162275:
>>>                          Processing:  { Ans: ,
>>>                          MgmtId: 37528005876872, via: 18, Ver: v1,
>>>         Flags: 10,
>>>
>>>         [{"Answer":{"result":true,"**_**___details":"success","wait":**
>>> 0}}__]
>>>
>>>
>>>                          }
>>>
>>>                          2013-04-16 16:27:14,323 DEBUG
>>>         [agent.transport.Request]
>>>                          (catalina-exec-9:null) Seq 18-1625162275:
>>>         Received:  {
>>>                          Ans: , MgmtId:
>>>                          37528005876872, via: 18, Ver: v1, Flags: 10, {
>>>         Answer } }
>>>                          2013-04-16 16:27:14,323 DEBUG
>>>                          [agent.manager.**____**AgentManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) Details from executing
>>> class
>>>
>>>         com.cloud.agent.api.**____**CreateStoragePoolCommand: success
>>>
>>>
>>>                          2013-04-16 16:27:14,323 DEBUG
>>>                          [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) In createPool Adding the
>>>         pool to
>>>                          each of the hosts
>>>                          2013-04-16 16:27:14,323 DEBUG
>>>                          [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) Adding pool ceph to
>>>  host 18
>>>                          2013-04-16 16:27:14,326 DEBUG
>>>         [agent.transport.Request]
>>>                          (catalina-exec-9:null) Seq 18-1625162276:
>>>         Sending  { Cmd
>>>                          , MgmtId:
>>>                          37528005876872, via: 18, Ver: v1, Flags: 100011,
>>>
>>>         [{"ModifyStoragePoolCommand":{**____**"add":true,"pool":{"id":**
>>> __218,__"**
>>>
>>>         uuid":"5924a2df-d658-3119-**__**__8aba-f90307683206","host":"**
>>> 10.____**
>>>
>>>         0.0.41","path":"cloudstack","***____*userInfo":":","port":**
>>> 6789,"__**
>>>
>>>         type":"RBD"},"localPath":"/**_**___mnt//3cf4f0e8-781d-39d8-**
>>> b81c-__*__*
>>>
>>>
>>>
>>>                          9896da212335","wait":0}}]
>>>                          }
>>>                          2013-04-16 16:27:14,411 DEBUG
>>>         [agent.transport.Request]
>>>                          (AgentManager-Handler-6:null) Seq 18-1625162276:
>>>                          Processing:  { Ans: ,
>>>                          MgmtId: 37528005876872, via: 18, Ver: v1,
>>>         Flags: 10,
>>>
>>>         [{"Answer":{"result":false,"****____details":"java.lang.**
>>>                          NullPointerException\n\tat
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> LibvirtStorageAdaptor.____**
>>>
>>>         createStoragePool(**____**LibvirtStorageAdaptor.java:**_**
>>> ___462)\n\tat
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> KVMStoragePoolManager.____**
>>>
>>>         createStoragePool(**____**KVMStoragePoolManager.java:57)**
>>> ____**\n\tat
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**___**_execute(
>>>                          **LibvirtComputingResource.___**
>>> _java:**2087)\n\tat
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**
>>>
>>>         executeRequest(**____**LibvirtComputingResource.java:**
>>> ____**1053)\n\tat
>>>
>>>         com.cloud.agent.Agent.**____**processRequest(Agent.java:518)**
>>> ____**\n\tat
>>>
>>>         com.cloud.agent.Agent$**____**AgentRequestHandler.doTask(**
>>>                          Agent.java:831)\n\tat
>>>
>>>         com.cloud.utils.nio.Task.run(***____*Task.java:83)\n\tat
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor.runWorker(***
>>> ____*
>>>                          ThreadPoolExecutor.java:1146)\**____**n\tat
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor$Worker.run(**
>>> ____**
>>>                          ThreadPoolExecutor.java:615)\***____*n\tat
>>>
>>>         java.lang.Thread.run(Thread.****____java:679)\n","wait":0}}] }
>>>
>>>
>>>
>>>                          2013-04-16 16:27:14,412 DEBUG
>>>         [agent.transport.Request]
>>>                          (catalina-exec-9:null) Seq 18-1625162276:
>>>         Received:  {
>>>                          Ans: , MgmtId:
>>>                          37528005876872, via: 18, Ver: v1, Flags: 10, {
>>>         Answer } }
>>>                          2013-04-16 16:27:14,412 DEBUG
>>>                          [agent.manager.**____**AgentManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) Details from executing
>>> class
>>>                          com.cloud.agent.api.**____**
>>> ModifyStoragePoolCommand:
>>>                          java.lang.NullPointerException
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> LibvirtStorageAdaptor.____**
>>>
>>>         createStoragePool(**____**LibvirtStorageAdaptor.java:**_**
>>> ___462)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> KVMStoragePoolManager.____**
>>>
>>>         createStoragePool(**____**KVMStoragePoolManager.java:57)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**___**_execute(
>>>                          **LibvirtComputingResource.___**_java:**2087)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**
>>>
>>>         executeRequest(**____**LibvirtComputingResource.java:**
>>> ____**1053)
>>>                                     at
>>>
>>>         com.cloud.agent.Agent.**____**processRequest(Agent.java:518)
>>>                                     at
>>>
>>>         com.cloud.agent.Agent$**____**AgentRequestHandler.doTask(**
>>>                          Agent.java:831)
>>>                                     at
>>>         com.cloud.utils.nio.Task.run(***____*Task.java:83)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor.runWorker(***
>>> ____*
>>>                          ThreadPoolExecutor.java:1146)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor$Worker.run(**
>>> ____**
>>>                          ThreadPoolExecutor.java:615)
>>>                                     at
>>>         java.lang.Thread.run(Thread.****____java:679)
>>>
>>>
>>>
>>>                          2013-04-16 16:27:14,451 WARN
>>>                            [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) Unable to establish a
>>>         connection
>>>                          between
>>>                          Host[-18-Routing] and Pool[218|RBD]
>>>
>>>         com.cloud.exception.**____**StorageUnavailableException:
>>>
>>>
>>>                          Resource
>>>
>>>                          [StoragePool:218]
>>>                          is unreachable: Unable establish connection
>>>         from storage
>>>                          head to storage
>>>                          pool 218 due to java.lang.NullPointerException
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> LibvirtStorageAdaptor.____**
>>>
>>>         createStoragePool(**____**LibvirtStorageAdaptor.java:**_**
>>> ___462)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_storage.**
>>> KVMStoragePoolManager.____**
>>>
>>>         createStoragePool(**____**KVMStoragePoolManager.java:57)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**___**_execute(
>>>                          **LibvirtComputingResource.___**_java:**2087)
>>>                                     at
>>>
>>>         com.cloud.hypervisor.kvm.**___**_resource.**____**
>>> LibvirtComputingResource.**
>>>
>>>         executeRequest(**____**LibvirtComputingResource.java:**
>>> ____**1053)
>>>                                     at
>>>
>>>         com.cloud.agent.Agent.**____**processRequest(Agent.java:518)
>>>                                     at
>>>
>>>         com.cloud.agent.Agent$**____**AgentRequestHandler.doTask(**
>>>                          Agent.java:831)
>>>                                     at
>>>         com.cloud.utils.nio.Task.run(***____*Task.java:83)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor.runWorker(***
>>> ____*
>>>                          ThreadPoolExecutor.java:1146)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor$Worker.run(**
>>> ____**
>>>                          ThreadPoolExecutor.java:615)
>>>                                     at
>>>         java.lang.Thread.run(Thread.****____java:679)
>>>
>>>                                     at
>>>
>>>         com.cloud.storage.**____**StorageManagerImpl.**____**
>>> connectHostToSharedPool(**
>>>                          StorageManagerImpl.java:1685)
>>>                                     at
>>>
>>>         com.cloud.storage.**____**StorageManagerImpl.createPool(**____**
>>>                          StorageManagerImpl.java:1450)
>>>                                     at
>>>
>>>         com.cloud.storage.**____**StorageManagerImpl.createPool(**____**
>>>                          StorageManagerImpl.java:215)
>>>                                     at
>>>
>>>         com.cloud.api.commands.**____**CreateStoragePoolCmd.execute(***
>>> ____*
>>>                          CreateStoragePoolCmd.java:120)
>>>                                     at
>>>
>>>         com.cloud.api.ApiDispatcher.****____dispatch(ApiDispatcher.**
>>> java:__**
>>>                          138)
>>>                                     at
>>>
>>>         com.cloud.api.ApiServer.**____**queueCommand(ApiServer.java:****
>>> ____543)
>>>                                     at
>>>
>>>         com.cloud.api.ApiServer.**____**handleRequest(ApiServer.java:***
>>> ____*422)
>>>                                     at
>>>
>>>         com.cloud.api.ApiServlet.**___**_processRequest(ApiServlet.**
>>>                          java:304)
>>>                                     at
>>>
>>>         com.cloud.api.ApiServlet.**___**_doGet(ApiServlet.java:63)
>>>                                     at
>>>         javax.servlet.http.**____**HttpServlet.service(**
>>>                          HttpServlet.java:617)
>>>                                     at
>>>         javax.servlet.http.**____**HttpServlet.service(**
>>>                          HttpServlet.java:717)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_ApplicationFilterChain.**____**
>>> internalDoFilter(**
>>>                          ApplicationFilterChain.java:****____290)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_ApplicationFilterChain.**____**
>>> doFilter(**
>>>                          ApplicationFilterChain.java:****____206)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_StandardWrapperValve.invoke(***
>>> *
>>>                          StandardWrapperValve.java:233)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_StandardContextValve.invoke(***
>>> *
>>>                          StandardContextValve.java:191)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_StandardHostValve.invoke(**
>>>                          StandardHostValve.java:127)
>>>                                     at
>>>
>>>         org.apache.catalina.valves.**_**___ErrorReportValve.invoke(**
>>>                          ErrorReportValve.java:102)
>>>                                     at
>>>
>>>         org.apache.catalina.valves.**_**___AccessLogValve.invoke(**
>>>                          AccessLogValve.java:555)
>>>                                     at
>>>
>>>         org.apache.catalina.core.**___**_StandardEngineValve.invoke(**
>>>                          StandardEngineValve.java:109)
>>>                                     at
>>>
>>>         org.apache.catalina.connector.**____**CoyoteAdapter.service(**
>>>                          CoyoteAdapter.java:298)
>>>                                     at
>>>
>>>         org.apache.coyote.http11.**___**_Http11NioProcessor.process(**
>>>                          Http11NioProcessor.java:889)
>>>                                     at
>>>
>>>         org.apache.coyote.http11.**___**_Http11NioProtocol$**____**
>>> Http11ConnectionHandler.**
>>>                          process(Http11NioProtocol.**__**__java:721)
>>>                                     at
>>>
>>>         org.apache.tomcat.util.net.**_**___NioEndpoint$**
>>> SocketProcessor.*__*
>>>                          run(NioEndpoint.java:2260)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor.runWorker(***
>>> ____*
>>>                          ThreadPoolExecutor.java:1110)
>>>                                     at
>>>
>>>         java.util.concurrent.**____**ThreadPoolExecutor$Worker.run(**
>>> ____**
>>>                          ThreadPoolExecutor.java:603)
>>>                                     at
>>>         java.lang.Thread.run(Thread.****____java:679)
>>>
>>>
>>>                          2013-04-16 16:27:14,452 WARN
>>>                            [cloud.storage.**____**StorageManagerImpl]
>>>
>>>
>>>
>>>                          (catalina-exec-9:null) No host can access
>>>         storage pool
>>>                          Pool[218|RBD] on
>>>                          cluster 5
>>>                          2013-04-16 16:27:14,504 WARN
>>>           [cloud.api.ApiDispatcher]
>>>                          (catalina-exec-9:null) class
>>>                          com.cloud.api.**____**ServerApiException :
>>> Failed
>>>
>>>
>>>                          to
>>>                          add storage pool
>>>                          2013-04-16 16:27:15,293 DEBUG
>>>                          [agent.manager.**____**AgentManagerImpl]
>>>
>>>
>>>
>>>                          (AgentManager-Handler-12:null) Ping from 18
>>>                          ^C
>>>                          [root@RDR02S02 management]#
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>         --
>>>         Guangjian
>>>
>>>
>>>
>>>
>>> --
>>> Guangjian
>>>
>>
>
>
> --
> Guangjian
>



-- 
Guangjian

Reply via email to