Hi Eugen.

Sorry for the delay in answering.

Just looked in the /var/log/ceph/ directory. It only contains the following
files (for example on node01):

#######
# ls -lart
total 3864
-rw------- 1 ceph ceph     904 ago 24 13:11 ceph.audit.log-20180829.xz
drwxr-xr-x 1 root root     898 ago 28 10:07 ..
-rw-r--r-- 1 ceph ceph  189464 ago 28 23:59 ceph-mon.node01.log-20180829.xz
-rw------- 1 ceph ceph   24360 ago 28 23:59 ceph.log-20180829.xz
-rw-r--r-- 1 ceph ceph   48584 ago 29 00:00 ceph-mgr.node01.log-20180829.xz
-rw------- 1 ceph ceph       0 ago 29 00:00 ceph.audit.log
drwxrws--T 1 ceph ceph     352 ago 29 00:00 .
-rw-r--r-- 1 ceph ceph 1908122 ago 29 12:46 ceph-mon.node01.log
-rw------- 1 ceph ceph  175229 ago 29 12:48 ceph.log
-rw-r--r-- 1 ceph ceph 1599920 ago 29 12:49 ceph-mgr.node01.log
#######

So, it only contains logs concerning the node itself (is it correct? sincer
node01 is also the master, I was expecting it to have logs from the other
too) and, moreover, no ceph-osd* files. Also, I'm looking the logs I have
available, and nothing "shines out" (sorry for my poor english) as a
possible error.

Any suggestion on how to proceed?

Thanks a lot in advance,

Jones


On Mon, Aug 27, 2018 at 5:29 AM Eugen Block <ebl...@nde.ag> wrote:

> Hi Jones,
>
> all ceph logs are in the directory /var/log/ceph/, each daemon has its
> own log file, e.g. OSD logs are named ceph-osd.*.
>
> I haven't tried it but I don't think SUSE Enterprise Storage deploys
> OSDs on partitioned disks. Is there a way to attach a second disk to
> the OSD nodes, maybe via USB or something?
>
> Although this thread is ceph related it is referring to a specific
> product, so I would recommend to post your question in the SUSE forum
> [1].
>
> Regards,
> Eugen
>
> [1] https://forums.suse.com/forumdisplay.php?99-SUSE-Enterprise-Storage
>
> Zitat von Jones de Andrade <johanne...@gmail.com>:
>
> > Hi Eugen.
> >
> > Thanks for the suggestion. I'll look for the logs (since it's our first
> > attempt with ceph, I'll have to discover where they are, but no problem).
> >
> > One thing called my attention on your response however:
> >
> > I haven't made myself clear, but one of the failures we encountered were
> > that the files now containing:
> >
> > node02:
> >    ----------
> >    storage:
> >        ----------
> >        osds:
> >            ----------
> >            /dev/sda4:
> >                ----------
> >                format:
> >                    bluestore
> >                standalone:
> >                    True
> >
> > Were originally empty, and we filled them by hand following a model found
> > elsewhere on the web. It was necessary, so that we could continue, but
> the
> > model indicated that, for example, it should have the path for /dev/sda
> > here, not /dev/sda4. We chosen to include the specific partition
> > identification because we won't have dedicated disks here, rather just
> the
> > very same partition as all disks were partitioned exactly the same.
> >
> > While that was enough for the procedure to continue at that point, now I
> > wonder if it was the right call and, if it indeed was, if it was done
> > properly.  As such, I wonder: what you mean by "wipe" the partition here?
> > /dev/sda4 is created, but is both empty and unmounted: Should a different
> > operation be performed on it, should I remove it first, should I have
> > written the files above with only /dev/sda as target?
> >
> > I know that probably I wouldn't run in this issues with dedicated discks,
> > but unfortunately that is absolutely not an option.
> >
> > Thanks a lot in advance for any comments and/or extra suggestions.
> >
> > Sincerely yours,
> >
> > Jones
> >
> > On Sat, Aug 25, 2018 at 5:46 PM Eugen Block <ebl...@nde.ag> wrote:
> >
> >> Hi,
> >>
> >> take a look into the logs, they should point you in the right direction.
> >> Since the deployment stage fails at the OSD level, start with the OSD
> >> logs. Something's not right with the disks/partitions, did you wipe
> >> the partition from previous attempts?
> >>
> >> Regards,
> >> Eugen
> >>
> >> Zitat von Jones de Andrade <johanne...@gmail.com>:
> >>
> >>> (Please forgive my previous email: I was using another message and
> >>> completely forget to update the subject)
> >>>
> >>> Hi all.
> >>>
> >>> I'm new to ceph, and after having serious problems in ceph stages 0, 1
> >> and
> >>> 2 that I could solve myself, now it seems that I have hit a wall harder
> >>> than my head. :)
> >>>
> >>> When I run salt-run state.orch ceph.stage.deploy, i monitor I see it
> >> going
> >>> up to here:
> >>>
> >>> #######
> >>> [14/71]   ceph.sysctl on
> >>>           node01....................................... ✓ (0.5s)
> >>>           node02........................................ ✓ (0.7s)
> >>>           node03....................................... ✓ (0.6s)
> >>>           node04......................................... ✓ (0.5s)
> >>>           node05....................................... ✓ (0.6s)
> >>>           node06.......................................... ✓ (0.5s)
> >>>
> >>> [15/71]   ceph.osd on
> >>>           node01...................................... ❌ (0.7s)
> >>>           node02........................................ ❌ (0.7s)
> >>>           node03....................................... ❌ (0.7s)
> >>>           node04......................................... ❌ (0.6s)
> >>>           node05....................................... ❌ (0.6s)
> >>>           node06.......................................... ❌ (0.7s)
> >>>
> >>> Ended stage: ceph.stage.deploy succeeded=14/71 failed=1/71 time=624.7s
> >>>
> >>> Failures summary:
> >>>
> >>> ceph.osd (/srv/salt/ceph/osd):
> >>>   node02:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node02 for cephdisks.list
> >>>   node03:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node03 for cephdisks.list
> >>>   node01:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node01 for cephdisks.list
> >>>   node04:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node04 for cephdisks.list
> >>>   node05:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node05 for cephdisks.list
> >>>   node06:
> >>>     deploy OSDs: Module function osd.deploy threw an exception.
> >> Exception:
> >>> Mine on node06 for cephdisks.list
> >>> #######
> >>>
> >>> Since this is a first attempt in 6 simple test machines, we are going
> to
> >>> put the mon, osds, etc, in all nodes at first. Only the master is left
> >> in a
> >>> single machine (node01) by now.
> >>>
> >>> As they are simple machines, they have a single hdd, which is
> partitioned
> >>> as follows (the hda4 partition is unmounted and left for the ceph
> >> system):
> >>>
> >>> ###########
> >>> # lsblk
> >>> NAME   MAJ:MIN RM   SIZE RO TYPE MOUNTPOINT
> >>> sda      8:0    0 465,8G  0 disk
> >>> ├─sda1   8:1    0   500M  0 part /boot/efi
> >>> ├─sda2   8:2    0    16G  0 part [SWAP]
> >>> ├─sda3   8:3    0  49,3G  0 part /
> >>> └─sda4   8:4    0   400G  0 part
> >>> sr0     11:0    1   3,7G  0 rom
> >>>
> >>> # salt -I 'roles:storage' cephdisks.list
> >>> node01:
> >>> node02:
> >>> node03:
> >>> node04:
> >>> node05:
> >>> node06:
> >>>
> >>> # salt -I 'roles:storage' pillar.get ceph
> >>> node02:
> >>>     ----------
> >>>     storage:
> >>>         ----------
> >>>         osds:
> >>>             ----------
> >>>             /dev/sda4:
> >>>                 ----------
> >>>                 format:
> >>>                     bluestore
> >>>                 standalone:
> >>>                     True
> >>> (and so on for all 6 machines)
> >>> ##########
> >>>
> >>> Finally and just in case, my policy.cfg file reads:
> >>>
> >>> #########
> >>> #cluster-unassigned/cluster/*.sls
> >>> cluster-ceph/cluster/*.sls
> >>> profile-default/cluster/*.sls
> >>> profile-default/stack/default/ceph/minions/*yml
> >>> config/stack/default/global.yml
> >>> config/stack/default/ceph/cluster.yml
> >>> role-master/cluster/node01.sls
> >>> role-admin/cluster/*.sls
> >>> role-mon/cluster/*.sls
> >>> role-mgr/cluster/*.sls
> >>> role-mds/cluster/*.sls
> >>> role-ganesha/cluster/*.sls
> >>> role-client-nfs/cluster/*.sls
> >>> role-client-cephfs/cluster/*.sls
> >>> ##########
> >>>
> >>> Please, could someone help me and shed some light on this issue?
> >>>
> >>> Thanks a lot in advance,
> >>>
> >>> Regasrds,
> >>>
> >>> Jones
> >>
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to