Thanks a lot Simon, this helps me to resolved the issue, it was the bug that you mentioned.
Best regards, *German* 2015-09-07 5:34 GMT-03:00 Simon Hallam <s...@pml.ac.uk>: > Hi German, > > > > This is what I’m running to redo an OSD as btrfs (not sure if this is the > exact error you’re getting): > > > > DISK_LETTER=( a b c d e f g h i j k l ) > > > > i=0 > > > > for OSD_NUM in {12..23}; do > > sudo /etc/init.d/ceph stop osd.${OSD_NUM} > > sudo umount /var/lib/ceph/osd/ceph-${OSD_NUM} > > sudo ceph auth del osd.${OSD_NUM} > > sudo ceph osd crush remove osd.${OSD_NUM} > > sudo ceph osd rm ${OSD_NUM} > > > > # recreate again > > sudo wipefs /dev/sd${DISK_LETTER[$i]}1 > > sudo dd if=/dev/zero of=/dev/sd${DISK_LETTER[$i]}1 bs=4k count=10000 > > sudo sgdisk --zap-all --clear -g /dev/sd${DISK_LETTER[$i]} > > sudo kpartx -dug /dev/sd${DISK_LETTER[$i]} > > sudo partprobe /dev/sd${DISK_LETTER[$i]} > > sudo dd if=/dev/zero of=/dev/sd${DISK_LETTER[$i]} bs=4k count=10000 > > sudo ceph-disk zap /dev/sd${DISK_LETTER[$i]} > > echo "" > > echo "ceph-deploy --overwrite-conf disk prepare --fs-type btrfs > ceph2:sd${DISK_LETTER[$i]}" > > echo "" > > read -p "Press [Enter] key to continue next disk after running the above > command on ceph1" > > i=$((i + 1)) > > done > > > > There appears to be an issue with zap not wiping the partitions correctly. > http://tracker.ceph.com/issues/6258 > > > > Yours seems slightly different though. Curious, what size disk are you > trying to use? > > > > Cheers, > > > > Simon > > > > *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf > Of *German Anders > *Sent:* 04 September 2015 19:53 > *To:* ceph-users > *Subject:* [ceph-users] ceph-deploy prepare btrfs osd error > > > > Any ideas? > > ceph@cephdeploy01:~/ceph-ib$ ceph-deploy osd prepare --fs-type btrfs > cibosd04:sdc > [ceph_deploy.conf][DEBUG ] found configuration file at: > /home/ceph/.cephdeploy.conf > [ceph_deploy.cli][INFO ] Invoked (1.5.28): /usr/bin/ceph-deploy osd > prepare --fs-type btrfs cibosd04:sdc > [ceph_deploy.cli][INFO ] ceph-deploy options: > [ceph_deploy.cli][INFO ] username : None > [ceph_deploy.cli][INFO ] disk : [('cibosd04', > '/dev/sdc', None)] > [ceph_deploy.cli][INFO ] dmcrypt : False > [ceph_deploy.cli][INFO ] verbose : False > [ceph_deploy.cli][INFO ] overwrite_conf : False > [ceph_deploy.cli][INFO ] subcommand : prepare > [ceph_deploy.cli][INFO ] dmcrypt_key_dir : > /etc/ceph/dmcrypt-keys > [ceph_deploy.cli][INFO ] quiet : False > [ceph_deploy.cli][INFO ] cd_conf : > <ceph_deploy.conf.cephdeploy.Conf instance at 0x7faf715a0bd8> > [ceph_deploy.cli][INFO ] cluster : ceph > [ceph_deploy.cli][INFO ] fs_type : btrfs > [ceph_deploy.cli][INFO ] func : <function osd > at 0x7faf71576938> > [ceph_deploy.cli][INFO ] ceph_conf : None > [ceph_deploy.cli][INFO ] default_release : False > [ceph_deploy.cli][INFO ] zap_disk : False > [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks cibosd04:/dev/sdc: > [cibosd04][DEBUG ] connection detected need for sudo > [cibosd04][DEBUG ] connected to host: cibosd04 > [cibosd04][DEBUG ] detect platform information from remote host > [cibosd04][DEBUG ] detect machine type > [cibosd04][DEBUG ] find the location of an executable > [ceph_deploy.osd][INFO ] Distro info: Ubuntu 14.04 trusty > [ceph_deploy.osd][DEBUG ] Deploying osd to cibosd04 > [cibosd04][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf > [cibosd04][INFO ] Running command: sudo udevadm trigger > --subsystem-match=block --action=add > [ceph_deploy.osd][DEBUG ] Preparing host cibosd04 disk /dev/sdc journal > None activate False > [cibosd04][INFO ] Running command: sudo ceph-disk -v prepare --cluster > ceph --fs-type btrfs -- /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --cluster=ceph --show-config-value=fsid > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_mkfs_options_btrfs > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_fs_mkfs_options_btrfs > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_mount_options_btrfs > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_fs_mount_options_btrfs > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --cluster=ceph --show-config-value=osd_journal_size > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_cryptsetup_parameters > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_dmcrypt_key_size > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_dmcrypt_type > [cibosd04][WARNIN] INFO:ceph-disk:Will colocate journal with data on > /dev/sdc > [cibosd04][WARNIN] DEBUG:ceph-disk:Creating journal partition num 2 size > 5120 on /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/sgdisk > --new=2:0:5120M --change-name=2:ceph journal > --partition-guid=2:e4d02c3f-0fd4-4270-a33f-15191cd86f1b > --typecode=2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 --mbrtogpt -- /dev/sdc > [cibosd04][DEBUG ] Creating new GPT entries. > [cibosd04][DEBUG ] The operation has completed successfully. > [cibosd04][WARNIN] DEBUG:ceph-disk:Calling partprobe on prepared device > /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/partprobe /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/udevadm settle > [cibosd04][WARNIN] DEBUG:ceph-disk:Journal is GPT partition > /dev/disk/by-partuuid/e4d02c3f-0fd4-4270-a33f-15191cd86f1b > [cibosd04][WARNIN] DEBUG:ceph-disk:Journal is GPT partition > /dev/disk/by-partuuid/e4d02c3f-0fd4-4270-a33f-15191cd86f1b > [cibosd04][WARNIN] DEBUG:ceph-disk:Creating osd partition on /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/sgdisk > --largest-new=1 --change-name=1:ceph data > --partition-guid=1:082e0de9-1d32-4502-ba78-4649cfaa0d83 > --typecode=1:89c57f98-2fe5-4dc0-89c1-f3ad0ceff2be -- /dev/sdc > [cibosd04][DEBUG ] The operation has completed successfully. > [cibosd04][WARNIN] DEBUG:ceph-disk:Calling partprobe on created device > /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/partprobe /dev/sdc > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/udevadm settle > [cibosd04][WARNIN] DEBUG:ceph-disk:Creating btrfs fs on /dev/sdc1 > [cibosd04][WARNIN] INFO:ceph-disk:Running command: /sbin/mkfs -t btrfs -m > single -l 32768 -n 32768 -- /dev/sdc1 > > > > > > *[cibosd04][DEBUG ] SMALL VOLUME: forcing mixed metadata/data groups > [cibosd04][WARNIN] Error: mixed metadata/data block groups require metadata > blocksizes equal to the sectorsize [cibosd04][WARNIN] ceph-disk: Error: > Command '['/sbin/mkfs', '-t', 'btrfs', '-m', 'single', '-l', '32768', '-n', > '32768', '--', '/dev/sdc1']' returned non-zero exit status 1 > [cibosd04][ERROR ] RuntimeError: command returned non-zero exit status: 1 > [ceph_deploy.osd][ERROR ] Failed to execute command: ceph-disk -v prepare > --cluster ceph --fs-type btrfs -- /dev/sdc [ceph_deploy][ERROR ] > GenericError: Failed to create 1 OSDs* > > Already try to do a gdisk /dev/sdc and delete the partitions, and run > again the command, but with the same error > > > > Thanks in advance, > > *German* > > Please visit our new website at www.pml.ac.uk and follow us on Twitter > @PlymouthMarine > > Winner of the Environment & Conservation category, the Charity Awards 2014. > > Plymouth Marine Laboratory (PML) is a company limited by guarantee registered > in England & Wales, company number 4178503. Registered Charity No. 1091222. > Registered Office: Prospect Place, The Hoe, Plymouth PL1 3DH, UK. > > This message is private and confidential. If you have received this message > in error, please notify the sender and remove it from your system. You are > reminded that e-mail communications are not secure and may contain viruses; > PML accepts no liability for any loss or damage which may be caused by > viruses. > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com