Hi,

It does not exist as below.

[root@rx200s7-07m ~]# rbd -p libvirt-pool info
5e5d9b40-270b-44af-9479-782175556c47
rbd: error opening image 5e5d9b40-270b-44af-9479-782175556c47: (2) No such
file or directory
2013-07-23 04:05:59.903162 7ff4d6b74760 -1 librbd::ImageCtx: error finding
header: (2) No such file or directory
[root@rx200s7-07m ~]#

> Can you try an install from an ISO to see if that works?
OK, I will try it tomorrow. (Now, it's 4AM on JST)


Thanks,
Satoshi Shimazaki


2013/7/23 Wido den Hollander <w...@widodh.nl>

> Hi,
>
>
> On 07/22/2013 08:55 PM, Satoshi Shimazaki wrote:
>
>> Hi Wido,
>>
>> Thank you for your comment.
>>
>>  What I see is "No such file or directory", so that RBD image does not
>>>
>> exist.
>>
>>> It seems like a copy didn't succeed but now CloudStack thinks that the
>>>
>> image does exist.
>>
>>> Does "libvirt-pool" have a RBD image with the name
>>>
>> 5e5d9b40-270b-44af-9479-**782175556c47 ?
>> No, it does not.
>>
>> There is no "5e5d9b40-270b-44af-9479-**782175556c47" as below.
>>
>>
> Oh, I meant this:
>
> $ rbd -p libvirt-pool info 5e5d9b40-270b-44af-9479-**782175556c47
>
>
>  virsh # pool-list
>> Name                 State      Autostart
>> ------------------------------**-----------
>> 3900a5bf-3362-392b-8bd0-**57b10ef47bb5 active     no
>> b39ca2cd-65ea-46d5-8a71-**c3a4ef95028e active     no
>> cd6520d6-bfc3-3537-9600-**7f044e11ddb1 active     no
>>
>> virsh # vol-list 3900a5bf-3362-392b-8bd0-**57b10ef47bb5
>> Name                 Path
>> ------------------------------**-----------
>> 6ff9719f-3e4d-4ff5-ab67-**154e30c936c2
>> libvirt-pool/6ff9719f-3e4d-**4ff5-ab67-154e30c936c2
>> 8555f35f-3ed8-436b-895a-**04e88e7327e0
>> libvirt-pool/8555f35f-3ed8-**436b-895a-04e88e7327e0
>> cd3688ab-e37b-4866-9ea7-**4051b670a323
>> libvirt-pool/cd3688ab-e37b-**4866-9ea7-4051b670a323
>>
>>
>>  What I see is "No such file or directory", so that RBD image does not
>>>
>> exist.
>>
>>> It seems like a copy didn't succeed but now CloudStack thinks that the
>>>
>> image does exist.
>> I agree, but I don't understand why it occured...
>>
>>
> I can't tell now. Can you try an install from an ISO to see if that works?
>
> And if that works, can you deploy a fresh template to see what the copy
> does?
>
> It will run qemu-img twice:
>
> 1. Secondary Storage -> RBD
> 2. RBD -> RBD
>
> In 4.2 the second step will be a RBD clone operation btw.
>
> Wido
>
>  - VMs(Root disk) on NFS : It works.
>> - VMs(Root disk) on RBD : It doesn't work
>> - Data Disk on RBD (attach to VMs on NFS) : It works.
>>
>> Is there any other points to be checked?
>>
>>
>> Thanks,
>> Satoshi Shimazaki
>>
>>
>>
>> 2013/7/23 Wido den Hollander <w...@widodh.nl>
>>
>>  Hi,
>>>
>>>
>>> On 07/22/2013 07:56 PM, Satoshi Shimazaki wrote:
>>>
>>>  Hi Wido,
>>>>
>>>> I'm in the project with Kimi and Nakajima-san.
>>>>
>>>> [root@rx200s7-07m ~]# ceph -v
>>>> ceph version 0.61.4 (****1669132fcfc27d0c0b5e5bb93ade59****d147e23404)
>>>>
>>>>
>>>> Same version is installed into all the hosts (KVM host and Ceph nodes).
>>>>
>>>> Here is KVM Agent log.
>>>> http://pastebin.com/5yG1uBuj
>>>> I had set the log level "DEBUG" and failed to create 2 instances
>>>> ,"RBDVM-shimazaki-1" and "RBDVM-shimazaki-2".
>>>>
>>>>
>>>>  What I see is "No such file or directory", so that RBD image does not
>>> exist.
>>>
>>> It seems like a copy didn't succeed but now CloudStack thinks that the
>>> image does exist.
>>>
>>> Does "libvirt-pool" have a RBD image with the name
>>> 5e5d9b40-270b-44af-9479-
>>> **782175556c47 ?
>>>
>>> Wido
>>>
>>>
>>>   Thanks,
>>>
>>>> Satoshi Shimazaki
>>>>
>>>>
>>>>
>>>>
>>>> 2013/7/23 Kimihiko Kitase <kimihiko.kit...@citrix.co.jp>
>>>>
>>>>   Hi Wido
>>>>
>>>>>
>>>>> Thanks for you comment.
>>>>>
>>>>> If we create vm on the NFS primary storage and mount additional disk on
>>>>> the RBD storage, it works fine.
>>>>> If we check vm from virt manager, there is no virtual disk. So we
>>>>> believe
>>>>> the problem should be vm configuration...
>>>>>
>>>>> We will check ceph version tomorrow.
>>>>>
>>>>> Thanks
>>>>> Kimi
>>>>>
>>>>> -----Original Message-----
>>>>> From: Wido den Hollander [mailto:w...@widodh.nl]
>>>>> Sent: Monday, July 22, 2013 11:43 PM
>>>>> To: dev@cloudstack.apache.org
>>>>> Subject: Re: Problem in adding Ceph RBD storage to CloudStack
>>>>>
>>>>> Hi,
>>>>>
>>>>> On 07/22/2013 02:25 PM, Kimihiko Kitase wrote:
>>>>>
>>>>>  Wido, Thank you very much.
>>>>>>
>>>>>> CloudStack: 4.1.0
>>>>>> QEMU: 1.5.50
>>>>>> Libvirt: 0.10.2
>>>>>>
>>>>>>
>>>>> What version of Ceph on the nodes?
>>>>>
>>>>> $ ceph -v
>>>>>
>>>>>
>>>>>  We will set "DEBUG" on the agent tomorrow. But the following is
>>>>>> command
>>>>>>
>>>>>>  CloudStack issue. We got this command at KVM host.
>>>>>
>>>>>
>>>>>> [root@rx200s7-07m ~]# ps -ef|grep 1517
>>>>>> root     16099     1 27 19:36 ?        00:00:12 /usr/libexec/qemu-kvm
>>>>>>
>>>>>>  -name i-2-1517-VM -S -M pc-i440fx-1.6 -enable-kvm -m 256 -smp
>>>>> 1,sockets=1,cores=1,threads=1 -uuid e67f1707-fe92-3426-978d-**
>>>>> 0441d5000d6a
>>>>> -no-user-config -nodefaults -chardev
>>>>> socket,id=charmonitor,path=/****var/lib/libvirt/qemu/i-2-1517-****
>>>>> VM.monitor,server,nowait
>>>>> -mon chardev=charmonitor,id=****monitor,mode=control -rtc base=utc
>>>>> -no-shutdown
>>>>> -boot dc -drive
>>>>> file=rbd:libvirt-pool/****cd3688ab-e37b-4866-9ea7-**
>>>>> 4051b670a323:id=libvirt:key=****AQC7OuZReMndFxAAY/**
>>>>> qUwLbvfod6EMvgVWU21g==:auth_****supported=cephx\;none:mon_**
>>>>> host=192.168.10.20\:6789,if=****none,id=drive-virtio-disk0,**
>>>>> format=raw,cache=none
>>>>> -device
>>>>> virtio-blk-pci,bus=pci.0,addr=****0x4,drive=drive-virtio-**disk0,**
>>>>> id=virtio-disk0
>>>>> -drive
>>>>> if=none,media=cdrom,id=drive-****ide0-1-0,readonly=on,format=****
>>>>> raw,cache=none
>>>>> -device ide-drive,bus=ide.1,unit=0,****drive=drive-ide0-1-0,id=ide0-**
>>>>> **1-0
>>>>> -netdev
>>>>> tap,fd=27,id=hostnet0,vhost=****on,vhostfd=29 -device
>>>>> virtio-net-pci,netdev=****hostnet0,id=net0,mac=02:00:0a:****b9:00:16
>>>>>
>>>>>    ,bus=pci.
>>>>> 0,addr=0x3 -chardev pty,id=charserial0 -device
>>>>> isa-serial,chardev=****charserial0,id=serial0 -usb -device
>>>>>
>>>>> usb-tablet,id=input0
>>>>> -vnc 0.0.0.0:3 -vga cirrus -device
>>>>> virtio-balloon-pci,id=****balloon0,bus=pci.0,addr=0x5
>>>>>
>>>>>
>>>>>
>>>>>>
>>>>>>  The argument to Qemu seems just fine, so I think the problem is not
>>>>> in
>>>>> CloudStack.
>>>>>
>>>>> Wido
>>>>>
>>>>>   Thanks
>>>>>
>>>>>> Kimi
>>>>>>
>>>>>> -----Original Message-----
>>>>>> From: Wido den Hollander [mailto:w...@widodh.nl]
>>>>>> Sent: Monday, July 22, 2013 7:47 PM
>>>>>> To: dev@cloudstack.apache.org
>>>>>> Subject: Re: Problem in adding Ceph RBD storage to CloudStack
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> On 07/22/2013 12:43 PM, Kimihiko Kitase wrote:
>>>>>>
>>>>>>  It seems secondary storage vm could copy template to primary storage
>>>>>>>
>>>>>>>  successfully, but created VM doesn't point this vol..
>>>>>>
>>>>>
>>>>>  If we create vm manually and add this vol as boot vol, it works fine..
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>  Which version of CloudStack are you using?
>>>>>>
>>>>>> What is the Qemu version running on your hypervisor and what libvirt
>>>>>>
>>>>>>  version?
>>>>>
>>>>>
>>>>>> If you set the logging level on the Agent to "DEBUG", does it show
>>>>>>
>>>>>>  deploying the VM with the correct XML parameters?
>>>>>
>>>>>
>>>>>> I haven't seen the things you are reporting.
>>>>>>
>>>>>> Wido
>>>>>>
>>>>>>   So it seems cloudstack cannot configure VM correctly in ceph rbd
>>>>>>
>>>>>>>
>>>>>>>  environment.
>>>>>>
>>>>>
>>>>>
>>>>>>  Any idea?
>>>>>>>
>>>>>>> Thanks
>>>>>>> Kimi
>>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: Kimihiko Kitase 
>>>>>>> [mailto:Kimihiko.Kitase@**citr**ix.co.jp<http://citrix.co.jp>
>>>>>>> <Kimihiko.Kitase@**citrix.co.jp <kimihiko.kit...@citrix.co.jp>>
>>>>>>> ]
>>>>>>> Sent: Monday, July 22, 2013 7:11 PM
>>>>>>> To: dev@cloudstack.apache.org
>>>>>>> Subject: RE: Problem in adding Ceph RBD storage to CloudStack
>>>>>>>
>>>>>>> Hello
>>>>>>>
>>>>>>> I am in the project with Nakajima san.
>>>>>>>
>>>>>>> We succeeded to add RBD storage to primary storage.
>>>>>>> But when we try to boot centos as user instance, it fail during
>>>>>>> system
>>>>>>>
>>>>>>>  logger process.
>>>>>>
>>>>>
>>>>>  It works fine when we boot centos using NFS storage.
>>>>>>
>>>>>>> It works fine when we boot centos using NFS storage and add
>>>>>>> additional
>>>>>>>
>>>>>>>  disk from RBD storage.
>>>>>>
>>>>>
>>>>>
>>>>>>  Do you have any idea to resolve this issue?
>>>>>>>
>>>>>>> Thanks
>>>>>>> Kimi
>>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: Takuma Nakajima 
>>>>>>> [mailto:penguin.trance.2716@****gmail.com<http://gmail.com>
>>>>>>> <penguin.trance.2716@**gmail.com <penguin.trance.2...@gmail.com>>
>>>>>>> ]
>>>>>>> Sent: Saturday, July 20, 2013 12:23 PM
>>>>>>> To: dev@cloudstack.apache.org
>>>>>>> Subject: Re: Problem in adding Ceph RBD storage to CloudStack
>>>>>>>
>>>>>>> I'm sorry but I forgot to tell you that the environment does not have
>>>>>>>
>>>>>>>  the internet connection.
>>>>>>
>>>>>
>>>>>  It is not allowed to make a direct connection to the internet because
>>>>>>
>>>>>>>
>>>>>>>  of the security policy.
>>>>>>
>>>>>
>>>>>
>>>>>>  Wido,
>>>>>>>
>>>>>>>  No, it works for me like a charm :)
>>>>>>>>
>>>>>>>> Could you set the Agent logging to DEBUG as well and show the output
>>>>>>>> of
>>>>>>>>
>>>>>>>>  that log? Maybe paste the log on pastebin.
>>>>>>>
>>>>>>>
>>>>>>>> I'm interested in the XMLs the Agent is feeding to libvirt when
>>>>>>>> adding
>>>>>>>>
>>>>>>>>  the RBD pool.
>>>>>>>
>>>>>>> I thought the new libvirt overwrites the old one, but actually both
>>>>>>>
>>>>>>>  libvirt (with RBD and without RBD) were installed to the system.
>>>>>> qemu
>>>>>>
>>>>> was
>>>>> installed from the package and so it might have the dependency to the
>>>>> libvirt installed from the package. After deleting the both libvirt
>>>>> installed from source and package, then installed it from rpm package
>>>>> with
>>>>> RBD support, RBD storage was registered to the CloudStack successfully.
>>>>>
>>>>>
>>>>>>  David,
>>>>>>>
>>>>>>>  Why not 6.4?
>>>>>>>>
>>>>>>>>
>>>>>>> Because of no internet connection, packages in the local mirror
>>>>>>>
>>>>>>>  repository may be old.
>>>>>>
>>>>>
>>>>>  I checked /etc/redhat-release and it showed the version is 6.3.
>>>>>>
>>>>>>>
>>>>>>> In current state, although the RBD storage was installed, system VMs
>>>>>>> won't start with "Unable to get vms
>>>>>>> org.libvirt.LibvirtException: Domain not found: no domain with
>>>>>>> matching uuid 'xxxxxxxx-xxxx-xxxx-xxxx-****xxxxxxxxxxxx'" error like
>>>>>>> http://mail-archives.apache.****org/mod_mbox/cloudstack-users/****
>>>>>>> 201303.mbox<http://mail-**archives.apache.org/mod_mbox/**
>>>>>>> cloudstack-users/201303.mbox<http://mail-archives.apache.org/mod_mbox/cloudstack-users/201303.mbox>
>>>>>>> >
>>>>>>> /
>>>>>>> %****3CD2EE6B3265AD864EB3EA4F5C670D****256F3546C1@EXMBX01L-CRP-03.**
>>>>>>> **
>>>>>>>
>>>>>>> webmdhealth.
>>>>>>> net%3E The uuid in the error message was not in the database of the
>>>>>>> management server nor in ceph storage node.
>>>>>>>
>>>>>>> I tried removing host from the CloudStack and cleaning up the
>>>>>>> computing
>>>>>>>
>>>>>>>  node, but it cannot be added again to the CloudStack.
>>>>>>
>>>>>
>>>>>  agent log says it attempted to connect to localhost:8250 though the
>>>>>>
>>>>>>>
>>>>>>>  management server address is set to 10.40.1.190 in global settings.
>>>>>>
>>>>>
>>>>>  management server log is here: http://pastebin.com/muGz73c0
>>>>>>
>>>>>>> (10.40.1.24 is the address of the computing node)
>>>>>>>
>>>>>>> Now the computing node is under rebuilding.
>>>>>>>
>>>>>>> Takuma Nakajima
>>>>>>>
>>>>>>> 2013/7/19 David Nalley <da...@gnsa.us>:
>>>>>>>
>>>>>>>  On Thu, Jul 18, 2013 at 12:09 PM, Takuma Nakajima
>>>>>>>> <penguin.trance.2...@gmail.com****> wrote:
>>>>>>>>
>>>>>>>>  Hi,
>>>>>>>>>
>>>>>>>>> I'm building a CloudStack 4.1 with Ceph RBD storage using RHEL 6.3
>>>>>>>>>
>>>>>>>>>  recently
>>>>>>>>
>>>>>>>
>>>>>>>  but it fails when adding RBD storage to primary storage.
>>>>>>>>
>>>>>>>>> Does anybody know about the problem?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> Why not 6.4?
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>

Reply via email to