Hi Michael,
you are right, my system is installed on disk sdc, and sda is the journal
disk to be shared.
This is the output of  partx -v /dev/sda, didn't see anything unusual:
device /dev/sda: start 0 size 117231408
gpt: 2 slices
# 1:      2048-  2099199 (  2097152 sectors,   1073 MB)
# 2:   2099200-  4196351 (  2097152 sectors,   1073 MB)
dos: 0 slices


2013/11/13 Michael <mich...@onlinefusion.co.uk>

>  Sorry, just spotted you're mounting on sdc. Can you chuck out a partx -v
> /dev/sda to see if there's anything odd about the data currently on there?
>
> -Michael
>
>
> On 12/11/2013 18:22, Michael wrote:
>
> As long as there's room on the SSD for the partitioner it'll just use the
> conf value for osd journal size to section it up as it adds OSD's (I
> generally use the "ceph-deploy osd create srv:data:journal e.g.
> srv-12:/dev/sdb:/dev/sde" format when adding disks).
> Does it being /dev/sda mean you're putting your journal onto an already
> partitioned and in use by the OS SSD?
>
> -Michael
>
> On 12/11/2013 18:09, Gruher, Joseph R wrote:
>
>  I didn’t think you could specify the journal in this manner (just
> pointing multiple OSDs on the same host all to journal /dev/sda).  Don’t
> you either need to partition the SSD and point each SSD to a separate
> partition, or format and mount the SSD and each OSD will use a unique file
> on the mount?  I’ve always created a separate partition on the SSD for each
> journal.
>
>
>
> Preparing cluster ceph disks ceph0:/dev/sdb:/dev/sda
> ceph0:/dev/sdd:/dev/sda ceph0:/dev/sde:/dev/sda ceph0:/dev/sdf:/dev/sda
> ceph0:/dev/sdg:/dev/sda ceph0:/dev/sdh:/dev/sda
>
>
>
> *From:* ceph-users-boun...@lists.ceph.com [
> mailto:ceph-users-boun...@lists.ceph.com<ceph-users-boun...@lists.ceph.com>]
> *On Behalf Of *Tim Zhang
> *Sent:* Tuesday, November 12, 2013 2:20 AM
> *To:* ceph-users@lists.ceph.com
> *Subject:* [ceph-users] ceph-deploy: osd creating hung with one ssd disk
> as shared journal
>
>
>
> Hi guys,
>
> I use ceph-deploy to manage my cluster, but I get failed while creating
> the OSD, the process seems to hang up at creating first osd. By the way,
> SELinux is disabled, and my ceph-disk is patched according to the page:
> http://www.spinics.net/lists/ceph-users/msg03258.html
>
> can you guys give me some advise?
>
> (1) the output of ceph-deploy is:
>
> Invoked (1.3.1): /usr/bin/ceph-deploy osd create ceph0:sdb:sda
> ceph0:sdd:sda ceph0:sde:sda ceph0:sdf:sda ceph0:sdg:sda ceph0:sdh:sda
> ceph1:sdb:sda ceph1:sdd:sda ceph1:sde:sda ceph1:sdf:sda ceph1:sdg:sda
> ceph1:sdh:sda ceph2:sdb:sda ceph2:sdd:sda ceph2:sde:sda ceph2:sdf:sda
> ceph2:sdg:sda ceph2:sdh:sda
>
> [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks
> ceph0:/dev/sdb:/dev/sda ceph0:/dev/sdd:/dev/sda ceph0:/dev/sde:/dev/sda
> ceph0:/dev/sdf:/dev/sda ceph0:/dev/sdg:/dev/sda ceph0:/dev/sdh:/dev/sda
> ceph1:/dev/sdb:/dev/sda ceph1:/dev/sdd:/dev/sda ceph1:/dev/sde:/dev/sda
> ceph1:/dev/sdf:/dev/sda ceph1:/dev/sdg:/dev/sda ceph1:/dev/sdh:/dev/sda
> ceph2:/dev/sdb:/dev/sda ceph2:/dev/sdd:/dev/sda ceph2:/dev/sde:/dev/sda
> ceph2:/dev/sdf:/dev/sda ceph2:/dev/sdg:/dev/sda ceph2:/dev/sdh:/dev/sda
>
> [ceph0][DEBUG ] connected to host: ceph0
>
> [ceph0][DEBUG ] detect platform information from remote host
>
> [ceph0][DEBUG ] detect machine type
>
> [ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final
>
> [ceph_deploy.osd][DEBUG ] Deploying osd to ceph0
>
> [ceph0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
>
> [ceph0][INFO  ] Running command: udevadm trigger --subsystem-match=block
> --action=add
>
> [ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdb journal
> /dev/sda activate True
>
> [ceph0][INFO  ] Running command: ceph-disk-prepare --fs-type xfs --cluster
> ceph -- /dev/sdb /dev/sda
>
> [ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal
> is not the same device as the osd data
>
> [ceph0][ERROR ] Warning: WARNING: the kernel failed to re-read the
> partition table on /dev/sda (Device or resource busy).  As a result, it may
> not reflect all of your changes until after reboot.
>
> [ceph0][ERROR ] BLKPG: Device or resource busy
>
> [ceph0][ERROR ] error adding partition 1
>
> [ceph0][DEBUG ] The operation has completed successfully.
>
> [ceph0][DEBUG ] The operation has completed successfully.
>
> [ceph0][DEBUG ] meta-data=/dev/sdb1              isize=2048   agcount=4,
> agsize=61047597 blks
>
> [ceph0][DEBUG ]          =                       sectsz=512   attr=2,
> projid32bit=0
>
> [ceph0][DEBUG ] data     =                       bsize=4096
> blocks=244190385, imaxpct=25
>
> [ceph0][DEBUG ]          =                       sunit=0      swidth=0 blks
>
> [ceph0][DEBUG ] naming   =version 2              bsize=4096   ascii-ci=0
>
> [ceph0][DEBUG ] log      =internal log           bsize=4096
> blocks=119233, version=2
>
> [ceph0][DEBUG ]          =                       sectsz=512   sunit=0
> blks, lazy-count=1
>
> [ceph0][DEBUG ] realtime =none                   extsz=4096   blocks=0,
> rtextents=0
>
> [ceph0][DEBUG ] The operation has completed successfully.
>
> [ceph0][INFO  ] Running command: udevadm trigger --subsystem-match=block
> --action=add
>
> [ceph_deploy.osd][DEBUG ] Host ceph0 is now ready for osd use.
>
> [ceph0][DEBUG ] connected to host: ceph0
>
> [ceph0][DEBUG ] detect platform information from remote host
>
> [ceph0][DEBUG ] detect machine type
>
> [ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final
>
> [ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdd journal
> /dev/sda activate True
>
> [ceph0][INFO  ] Running command: ceph-disk-prepare --fs-type xfs --cluster
> ceph -- /dev/sdd /dev/sda
>
> [ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal
> is not the same device as the osd data
>
>
>
> 2 the mount system for that osd shows:
>
> [root@host ~]# mount -l
>
> /dev/sdc1 on / type ext4 (rw)
>
> proc on /proc type proc (rw)
>
> sysfs on /sys type sysfs (rw)
>
> devpts on /dev/pts type devpts (rw,gid=5,mode=620)
>
> tmpfs on /dev/shm type tmpfs (rw)
>
> none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
>
> /dev/sdb1 on /var/lib/ceph/tmp/mnt.6D02EM type xfs (rw,noatime)
>
>
>
> 3 my testbed information is:
>
> os: centos 6.4 Final
>
> ceph: dumpling 67.4
>
> three hosts: ceph0 ceph1 ceph2
>
> each host have 3 disk sharing one ssd disk as journal
>
>
>
> 4 my ceph config is as this:
>
> osd journal size = 9500
>
> ;osd mkfs type = xfs
>
> ;auth supported = none
>
> auth_cluster_required = none
>
> auth_service_required = none
>
> auth_client_required = none
>
> public_network = 172.18.11.0/24
>
> cluster_network = 10.10.11.0/24
>
> osd pool default size = 3
>
> ms nocrc = true
>
> osd op threads = 4
>
> filestore op threads = 0
>
> mon sync fs threshold = 0
>
> osd pool default pg num = 100
>
> osd pool default pgp num = 100
>
>
>
> 5 the output of ceph0 running command: pe -ef|grep ceph
>
> [root@ceph0 ~]# ps -ef|grep ceph
>
> root     13922     1  0 05:59 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sdb1 sdb
>
> root     14059 13922  0 05:59 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate /dev/sdb1
>
> root     14090     1  0 05:59 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sda1 sda
>
> root     14107 14090  0 05:59 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate-journal /dev/sda1
>
> root     14220 14059  0 05:59 ?        00:00:00 /usr/bin/ceph-osd
> --cluster ceph --mkfs --mkkey -i 1 --monmap
> /var/lib/ceph/tmp/mnt.6D02EM/activate.monmap --osd-data
> /var/lib/ceph/tmp/mnt.6D02EM --osd-journal
> /var/lib/ceph/tmp/mnt.6D02EM/journal --osd-uuid
> 68381525-92ae-4e35-bf7f-aa34e31f2fae --keyring
> /var/lib/ceph/tmp/mnt.6D02EM/keyring
>
> root     14280     1  0 06:02 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sdb1 sdb
>
> root     14305 14280  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate /dev/sdb1
>
> root     14337     1  0 06:02 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sda1 sda
>
> root     14344     1  0 06:02 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 2 sda2 sda
>
> root     14366 14337  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate-journal /dev/sda1
>
> root     14371 14344  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate-journal /dev/sda2
>
> root     14409     1  0 06:05 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sdb1 sdb
>
> root     14430 14409  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate /dev/sdb1
>
> root     14444     1  0 06:05 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 1 sda1 sda
>
> root     14457     1  0 06:05 ?        00:00:00 /bin/sh
> /usr/sbin/ceph-disk-udev 2 sda2 sda
>
> root     14468 14444  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate-journal /dev/sda1
>
> root     14483 14457  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk
> -v activate-journal /dev/sda2
>
> root     14601 13254  0 06:18 pts/2    00:00:00 grep ceph
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to