I am also using centos7.x. /usr/lib/udev/rules.d/ should be fine. If not,
one can always symlink to /etc/udev/rules.d/.

On 19 November 2015 at 14:13, David Riedl <david.ri...@wingcon.com> wrote:

> Thanks for the fix!
> Two questions though:
> Is that the right place for the udev rule? I have CentOS 7. The folder
> exists, but all the other udev rules are in /usr/lib/udev/rules.d/.
> Can I just create a new file named "89-ceph-journal.rules"  in the
> /usr/lib/udev/rules.d/ folder?
>
>
> Regards
>
> David
>
>
> On 19.11.2015 14:02, Mykola Dvornik wrote:
>
> cat /etc/udev/rules.d/89-ceph-journal.rules
>
> KERNEL=="sdd?" SUBSYSTEM=="block" OWNER="ceph" GROUP="disk" MODE="0660"
>
> On 19 November 2015 at 13:54, Mykola <mykola.dvor...@gmail.com> wrote:
>
>> I am afraid one would need an udev rule to make it persistent.
>>
>>
>>
>> Sent from Outlook Mail <http://go.microsoft.com/fwlink/?LinkId=550987>
>> for Windows 10 phone
>>
>>
>>
>>
>> *From: *David Riedl <david.ri...@wingcon.com>
>> *Sent: *Thursday, November 19, 2015 1:42 PM
>> *To: *ceph-us...@ceph.com
>> *Subject: *Re: [ceph-users] Can't activate osd in infernalis
>>
>>
>>
>> I fixed the issue and opened a ticket on the ceph-deploy bug tracker
>>
>> http://tracker.ceph.com/issues/13833
>>
>>
>>
>> tl;dr:
>>
>> change permission of the ssd journal partition with
>>
>> chown ceph:ceph /dev/sdd1
>>
>>
>>
>> On 19.11.2015 11:38, David Riedl wrote:
>>
>> > Hi everyone.
>>
>> > I updated one of my hammer osd nodes to infernalis today.
>>
>> > After many problems with the upgrading process of the running OSDs, I
>>
>> > decided to wipe them and start anew.
>>
>> > I reinstalled all packages and deleted all partitions on the OSDs and
>>
>> > the SSD journal drive.
>>
>> > I zapped the disks with ceph-deploy and also prepared them with
>>
>> > ceph-deploy.
>>
>> > Selinux state is enabled (disabling it didn't help though).
>>
>> >
>>
>> > After executing "ceph-deploy osd activate ceph01:/dev/sda1:/dev/sdd1"
>>
>> > I get the following error message from ceph-deploy:
>>
>> >
>>
>> >
>>
>> > [ceph01][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph
>>
>> > --cluster ceph --name client.bootstrap-osd --keyring
>>
>> > /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu/activate.monmap
>>
>> > [ceph01][WARNIN] 2015-11-19 11:22:53.974765 7f1a06852700  0 --
>>
>> > :/3225863658 >> 10.20.60.10:6789/0 pipe(0x7f19f8062590 sd=4 :0 s=1
>>
>> > pgs=0 cs=0 l=1 c=0x7f19f805c1b0).fault
>>
>> > [ceph01][WARNIN] got monmap epoch 16
>>
>> > [ceph01][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd
>>
>> > --cluster ceph --mkfs --mkkey -i 0 --monmap
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu/activate.monmap --osd-data
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu --osd-journal
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu/journal --osd-uuid
>>
>> > de162e24-16b6-4796-b6b9-774fdb8ec234 --keyring
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu/keyring --setuser ceph --setgroup ceph
>>
>> > [ceph01][WARNIN] 2015-11-19 11:22:57.237096 7fb458bb7900 -1
>>
>> > filestore(/var/lib/ceph/tmp/mnt.pmHRuu) mkjournal error creating
>>
>> > journal on /var/lib/ceph/tmp/mnt.pmHRuu/journal: (13) Permission denied
>>
>> > [ceph01][WARNIN] 2015-11-19 11:22:57.237118 7fb458bb7900 -1 OSD::mkfs:
>>
>> > ObjectStore::mkfs failed with error -13
>>
>> > [ceph01][WARNIN] 2015-11-19 11:22:57.237157 7fb458bb7900 -1  ** ERROR:
>>
>> > error creating empty object store in /var/lib/ceph/tmp/mnt.pmHRuu:
>>
>> > (13) Permission denied
>>
>> > [ceph01][WARNIN] ERROR:ceph-disk:Failed to activate
>>
>> > [ceph01][WARNIN] DEBUG:ceph-disk:Unmounting /var/lib/ceph/tmp/mnt.pmHRuu
>>
>> > [ceph01][WARNIN] INFO:ceph-disk:Running command: /bin/umount --
>>
>> > /var/lib/ceph/tmp/mnt.pmHRuu
>>
>> > [ceph01][WARNIN] Traceback (most recent call last):
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 3576, in <module>
>>
>> > [ceph01][WARNIN]     main(sys.argv[1:])
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 3530, in main
>>
>> > [ceph01][WARNIN]     args.func(args)
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 2424, in
>>
>> > main_activate
>>
>> > [ceph01][WARNIN]     dmcrypt_key_dir=args.dmcrypt_key_dir,
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 2197, in
>>
>> > mount_activate
>>
>> > [ceph01][WARNIN]     (osd_id, cluster) = activate(path,
>>
>> > activate_key_template, init)
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 2360, in activate
>>
>> > [ceph01][WARNIN]     keyring=keyring,
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 1950, in mkfs
>>
>> > [ceph01][WARNIN]     '--setgroup', get_ceph_user(),
>>
>> > [ceph01][WARNIN]   File "/usr/sbin/ceph-disk", line 349, in
>>
>> > command_check_call
>>
>> > [ceph01][WARNIN]     return subprocess.check_call(arguments)
>>
>> > [ceph01][WARNIN]   File "/usr/lib64/python2.7/subprocess.py", line
>>
>> > 542, in check_call
>>
>> > [ceph01][WARNIN]     raise CalledProcessError(retcode, cmd)
>>
>> > [ceph01][WARNIN] subprocess.CalledProcessError: Command
>>
>> > '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', '-i',
>>
>> > '0', '--monmap', '/var/lib/ceph/tmp/mnt.pmHRuu/activate.monmap',
>>
>> > '--osd-data', '/var/lib/ceph/tmp/mnt.pmHRuu', '--osd-journal',
>>
>> > '/var/lib/ceph/tmp/mnt.pmHRuu/journal', '--osd-uuid',
>>
>> > 'de162e24-16b6-4796-b6b9-774fdb8ec234', '--keyring',
>>
>> > '/var/lib/ceph/tmp/mnt.pmHRuu/keyring', '--setuser', 'ceph',
>>
>> > '--setgroup', 'ceph']' returned non-zero exit status 1
>>
>> > [ceph01][ERROR ] RuntimeError: command returned non-zero exit status: 1
>>
>> > [ceph_deploy][ERROR ] RuntimeError: Failed to execute command:
>>
>> > ceph-disk -v activate --mark-init systemd --mount /dev/sda1
>>
>> >
>>
>> > The output of ls -lahn in /var/lib/ceph/ is
>>
>> >
>>
>> > drwxr-x---.  9 167 167 4,0K 19. Nov 10:32 .
>>
>> > drwxr-xr-x. 28   0   0 4,0K 19. Nov 11:14 ..
>>
>> > drwxr-x---.  2 167 167    6 10. Nov 13:06 bootstrap-mds
>>
>> > drwxr-x---.  2 167 167   25 19. Nov 10:48 bootstrap-osd
>>
>> > drwxr-x---.  2 167 167    6 10. Nov 13:06 bootstrap-rgw
>>
>> > drwxr-x---.  2 167 167    6 10. Nov 13:06 mds
>>
>> > drwxr-x---.  2 167 167    6 10. Nov 13:06 mon
>>
>> > drwxr-x---.  2 167 167    6 10. Nov 13:06 osd
>>
>> > drwxr-x---.  2 167 167   65 19. Nov 11:22 tmp
>>
>> >
>>
>> >
>>
>> > I hope someone can help me, I am really lost right now.
>>
>> >
>>
>>
>>
>> --
>>
>> Mit freundlichen Grüßen
>>
>>
>>
>> David Riedl
>>
>>
>>
>>
>>
>>
>>
>> WINGcon GmbH Wireless New Generation - Consulting & Solutions
>>
>>
>>
>> Phone: +49 (0) 7543 9661 - 26 <%2B49%20%280%29%207543%209661%20-%2026>
>>
>> E-Mail: <david.ri...@wingcon.com>david.ri...@wingcon.com
>>
>> Web: http://www.wingcon.com
>>
>>
>>
>> Sitz der Gesellschaft: Langenargen
>>
>> Registergericht: ULM, HRB 632019
>>
>> USt-Id.: DE232931635, WEEE-Id.: DE74015979
>>
>> Geschäftsführer: Norbert Schäfer, Fritz R. Paul
>>
>>
>>
>> _______________________________________________
>>
>> ceph-users mailing list
>>
>> ceph-users@lists.ceph.com
>>
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>
>
>
> --
>  Mykola
>
>
> --
> Mit freundlichen Grüßen
>
> David Riedl
>
>
>
> WINGcon GmbH Wireless New Generation - Consulting & Solutions
>
> Phone: +49 (0) 7543 9661 - 26
> E-Mail: david.ri...@wingcon.com
> Web: http://www.wingcon.com
>
> Sitz der Gesellschaft: Langenargen
> Registergericht: ULM, HRB 632019
> USt-Id.: DE232931635, WEEE-Id.: DE74015979
> Geschäftsführer: Norbert Schäfer, Fritz R. Paul
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
 Mykola
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to