Hi,
  I followed this documentation and didn't specify any CRUSH settings.

http://ceph.com/docs/next/rbd/rbd-openstack/

--
Jean-Sébastien Frerot
jsfre...@egliseespoir.com


2013/10/10 Gregory Farnum <g...@inktank.com>

> Okay. As a quick guess you probably used a CRUSH placement option with
> your new pools that wasn't supported by the old kernel, although it
> might have been something else.
>
> I suspect that you'll find FUSE works better for you anyway as long as
> you can use it — faster updates from us to you. ;)
> -Greg
> Software Engineer #42 @ http://inktank.com | http://ceph.com
>
>
> On Thu, Oct 10, 2013 at 10:53 AM, Frerot, Jean-Sébastien
> <jsfre...@egliseespoir.com> wrote:
> > Hi,
> >   Thx for your reply :)
> >
> > kernel: Linux compute01 3.8.0-31-generic #46-Ubuntu SMP Tue Sep 10
> 20:03:44
> > UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
> >
> > So yes I'm using cephfs and was also using rdb at the same time using
> > different pools. My ceph fs was setup 3 months ago and I upgraded it a
> > couple of days ago. I move VM images from rdb to the cephfs by copying
> the
> > file from rdb to local FS then to cephfs.
> >
> > I create pools like this:
> > ceph osd pool create volumes 128
> > ceph osd pool create images 128
> > ceph osd pool create live_migration 128
> >
> > Yes I had checked dmesg but didn't find anything relevant.
> >
> > However, as a last resort I decided to mount my FS using fuse. And it
> works
> > like a charm. So for now I'm sticking with fuse :)
> >
> > Let me know if you want me to do some explicit testing. It may take some
> > time for me to do them as I'm using ceph but I can manage to have some
> time
> > for maintenances.
> >
> > Regards,
> >
> >
> > --
> > Jean-Sébastien Frerot
> > jsfre...@egliseespoir.com
> >
> >
> > 2013/10/10 Gregory Farnum <g...@inktank.com>
> >>
> >> (Sorry for the delayed response, this was in my spam folder!)
> >>
> >> Has this issue persisted? Are you using the stock 13.04 kernel?
> >>
> >> Can you describe your setup a little more clearly? It sounds like
> >> maybe you're using CephFS now and were using rbd before; is that
> >> right? What data did you move, when, and how did you set up your
> >> CephFS to use the pools?
> >> The socket errors are often a slightly spammy notification that the
> >> socket isn't in use but has shut down; here they look to be an
> >> indicator of something actually gone wrong — perhaps you've
> >> inadvertently activated features incompatible with your kernel client,
> >> but let's see what's going on more before we jump to that conclusion.
> >> have you checked dmesg for anything else at those points?
> >> -Greg
> >> Software Engineer #42 @ http://inktank.com | http://ceph.com
> >>
> >> On Sat, Oct 5, 2013 at 6:42 PM, Frerot, Jean-Sébastien
> >> <jsfre...@egliseespoir.com> wrote:
> >> > Hi,
> >> >   I have a ceph cluster running with 3 physical servers,
> >> >
> >> > Here is how my setup is configured
> >> > server1: mon, osd, mds
> >> > server2: mon, osd, mds
> >> > server3: mon
> >> > OS ubuntu 13.04
> >> > ceph version: 0.67.4-1raring (recentrly upgrade to see if my problem
> >> > still
> >> > persisted with the new version)
> >> >
> >> > So I was running version CUTTLEFISH until yesterday. And I was using
> >> > ceph
> >> > with openstack (using rdb) but I simplified my setup and removed
> >> > openstack
> >> > to simply use kvm with virtmanager.
> >> >
> >> > So I created a new pool to be able to do live migration of kvm
> instances
> >> > #ceph osd lspools
> >> > 0 data,1 metadata,2 rbd,3 volumes,4 images,6 live_migration,
> >> >
> >> > I've been running VMs for some days without problems, but then I
> notice
> >> > that
> >> > I couldn't use the full disk size of my first VM (web01 which was 160G
> >> > big
> >> > originaly) but now is only 119G stored in ceph. I also have a windows
> >> > instance running on a 300G raw file located in ceph too. So trying to
> >> > fix
> >> > the issue I decided to do a local backup of my file in cause something
> >> > goes
> >> > wrong and guess what, i wasn't able to copy the file from ceph to my
> >> > local
> >> > drive. The moment I tried to do that "cp live_migration/web01 /mnt/"
> the
> >> > OS
> >> > hangs, and syslog show this >30 lines/s:
> >> >
> >> > Oct  5 15:25:45 server2 kernel: [ 8773.432358] libceph: osd1
> >> > 192.168.0.131:6803 socket error on read
> >> >
> >> > i couldn't kill my cp neither normally reboot my server. So I had to
> >> > reset
> >> > it.
> >> >
> >> > I tried to copy my other file "win2012" also stored in the ceph
> cluster
> >> > and
> >> > get the same issue and now I can't read anything from it nor start my
> VM
> >> > again
> >> >
> >> > [root@server1 ~]# ceph status
> >> >   cluster 50dc0404-c081-4c43-ac3f-872ba5494bd7
> >> >    health HEALTH_OK
> >> >    monmap e4: 3 mons at
> >> >
> >> > {server1=
> 192.168.0.130:6789/0,server2=192.168.0.131:6789/0,server3=192.168.0.132:6789/0
> },
> >> > election epoch 120, quorum 0,1,2 server1,server2,server3
> >> >    osdmap e275: 2 osds: 2 up, 2 in
> >> >     pgmap v1508209: 576 pgs: 576 active+clean; 108 GB data, 214 GB
> used,
> >> > 785
> >> > GB / 999 GB avail
> >> >    mdsmap e181: 1/1/1 up {0=server2=up:active}, 1 up:standby
> >> >
> >> > I mount the FS with fstab like this:
> >> > 192.168.0.131:6789,192.168.0.130:6789:/live_migration
> /var/lib/instances
> >> > ceph name=live_migration,secret=mysecret==,noatime 0 2
> >> >
> >> > I get this log in ceph-osd.0.log as spammy as "socket error on read"
> >> > error i
> >> > get in syslog
> >> > 2013-10-05 23:07:23.586807 7f24731cc700  0 --
> 192.168.0.130:6801/19182
> >> > >>
> >> > 192.168.0.130:0/4212596483 pipe(0x128d8500 sd=115 :6801 s=0 pgs=0
> cs=0
> >> > l=0
> >> > c=0x14ac09a0).accept peer addr is rea
> >> > lly 192.168.0.130:0/4212596483 (socket is 192.168.0.130:35078/0)
> >> >
> >> > other infos:
> >> > df -h
> >> > /dev/mapper/server1--vg-ceph                         500G  108G  393G
> >> > 22%
> >> > /opt/data/ceph
> >> > 192.168.0.131:6789,192.168.0.130:6789:/live_migration 1000G  215G
>  786G
> >> > 22%
> >> > /var/lib/instances
> >> > ...
> >> >
> >> > mount
> >> > /dev/mapper/server1--vg-ceph on /opt/data/ceph type xfs (rw,noatime)
> >> > 192.168.0.131:6789,192.168.0.130:6789:/live_migration on
> >> > /var/lib/instances
> >> > type ceph (name=live_migration,key=client.live_migration)
> >> > ...
> >> >
> >> >
> >> > How can I recover from this ?
> >> >
> >> > Thank you,
> >> > --
> >> > Jean-Sébastien Frerot
> >> > jsfre...@egliseespoir.com
> >> >
> >> > _______________________________________________
> >> > ceph-users mailing list
> >> > ceph-users@lists.ceph.com
> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >> >
> >
> >
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to