Hi,

I’m coming at this from the position of a newbie to Ceph.  I had some 
experience of it as part of Proxmox, but not as a standalone solution.

I really don’t care whether Ceph is contained or not, I don’t have the depth of 
knowledge or experience to argue it either way.  I can see that containers may 
well offer a more consistent deployment scenario with fewer dependencies on the 
external host OS.  Upgrades/patches to the host OS may not impact the container 
deployment etc., with the two systems not held in any lock-step.

The challenge for me hasn’t been Ceph its self. Ceph has worked brilliantly, I 
have a fully resilient architecture split between two active datacentres and my 
storage can survive up-to 50% node/OSD hardware failure.

No, the challenge has been documentation.  I’ve run off down multiple rabbit 
holes trying to find solutions to problems or just background information.  
I’ve been tripped up by not spotting the Ceph documentation was “v: latest” 
rather than “v: octopus”...so features didn’t exist or commands were structured 
slightly differently.

Also just not being obvious whether the bit of documentation I was looking at 
related to a native Ceph package deployment or a container one.  Plus you get 
the Ceph/Suse/Redhat/Proxmox/IBM etc..etc.. flavour answer depending on which 
Google link you click.  Yes I know, its part of the joy of working with open 
source....but still, not what you need when I chunk of infrastructure has 
failed and you don’t know why.

I’m truly in awe of what the Ceph community has produced and is planning for 
the future, so don’t think I’m any kind of hater.

My biggest request is for the documentation to take on some restructuring.  
Keep the different deployment methods documented separately, yes an intro 
covering the various options and recommendations is great, but then keep it 
entirely discreet.

Then when a feature/function is documented, make it clear if this applies to 
packaged or container deployment etc...  e.g. Zabbix (we use Zabbix)....lovely 
documentation on how to integrate Ceph and Zabbix....until you finally find out 
its not supported with containers....via a forum and an RFE/Bug entry.

And thank you to all the support in the community, REALLY appreciated.

Best,

Andrew


Sent from Mail<https://go.microsoft.com/fwlink/?LinkId=550986> for Windows

From: Erik Lindahl<mailto:erik.lind...@gmail.com>
Sent: 17 August 2021 16:01
To: Marc<mailto:m...@f1-outsourcing.eu>
Cc: Nico Schottelius<mailto:nico.schottel...@ungleich.ch>; Kai 
Börnert<mailto:kai.boern...@posteo.de>; ceph-users<mailto:ceph-users@ceph.io>
Subject: [ceph-users] Re: Why you might want packages not containers for Ceph 
deployments

Hi,

Whether containers are good or not is a separate discussion where I suspect 
there won't be consensus in the near future.

However, after just having looked at the documentation again, my main point 
would be that when a major stable open source project recommends a specific 
installation method (=cephadm) first in the "getting started" guide, users are 
going to expect that's the alternative things are documented for, which isn't 
quite the case for cephadm (yet).

Most users will probably accept either solution as long as there is ONE clear & 
well-documented way of working with ceph - but the current setup of even having 
the simple (?) getting started guide talk about at least three different ways 
without clearly separating their documentation seems like a guarantee for 
long-term confusion and higher entry barriers for new users, which I assume is 
the opposite of the goal of cephadm!

Cheers,

Erik


Erik Lindahl <erik.lind...@scilifelab.se>
Professor of Biophysics
Science for Life Laboratory
Stockholm University & KTH
Office (SciLifeLab): +46 8 524 81567
Cell (Sweden): +46 73 4618050
Cell (US): +1 (650) 924 7674



> On 17 Aug 2021, at 16:29, Marc <m...@f1-outsourcing.eu> wrote:
>
> 
>>
>>
>> Again, this is meant as hopefully constructive feedback rather than
>> complaints, but the feeling a get after having had fairly smooth
>> operations with raw packages (including fixing previous bugs leading to
>> severe crashes) and lately grinding our teeth a bit over cephadm is that
>> it has helped automated a bunch of stuff that wasn't particularly
>> difficult (it's nice to issue an update with a single command, but it
>> works perfectly fine manually too) at the cost of making it WAY more
>> difficult to fix things (not to mention simply get information about the
>> cluster) when we have problems - and in the long run that's not a trade-
>> off I'm entirely happy with :-)
>>
>
> Everyone can only agree to keeping things simple. I honestly do not even know 
> why you want to try cephadm. The containerized solution was developed to 
> replace ceph deploy, ceph ansible etc. as a solution to make ceph 
> installation for new users easier. That is exactly the reason (imho) why you 
> should not use the containerized environment. Because a containerized 
> environment has not as primaray task being an easy deployment tool. And 
> because the focus is on easy deployment, the real characteristics of the 
> containerized environment are being ignored during this development. Such as, 
> you must be out of your mind to create a depency between ceph-osd/msd/mon/all 
> and dockerd.
>
> 10 years(?) ago the people of mesos thought the docker containerizer was 
> 'flacky' and created their own more stable containerizer. And still today, 
> containers are being killed if dockerd is terminated. What some users had to 
> learn the hard way, as recently posted here.
>
> Today's container solutions are not on the level where you can say, you 
> require absolutely no knowledge to fix issues. So that means you would always 
> require knowledge of the container solution + ceph to troubleshoot. And that 
> is of course more knowledge, than just knowing ceph.
>
> I would not be surprised if cephadm ends up like ceph deploy/ansible.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to