>> It is nonetheless risky.  The wrong sequence of cascading events, of 
>> overlapping failures and you may lose data.  
> 
> Our setup is with 3/2.  size=3 seems much safer than 2.

Indeed that is the default for replicated pools.  Additional replicas exhibit 
diminishing returns in most cases at high cost.

> 
>>> 2-) Move the filesystem metadata pools to use at least SSD only.
>>> 
>> Absolutely.  The CephFS docs suggest using size=4 for the MD pool.
>> 
> 
> Hmm..  I don’t remember reading that anywhere, but it makes sense.  

https://docs.ceph.com/en/quincy/cephfs/createfs/#creating-pools

We recommend configuring at least 3 replicas for the metadata pool, as data 
loss in this pool can render the entire file system inaccessible. Configuring 4 
would not be extreme, especially since the metadata pool’s capacity 
requirements are quite modest.


> 
> Thanks!
> 
> George
> 
> 
>> 
>>> 
>>> 3-) Increase server and client cache.
>>> Here I left it like this:
>>> osd_memory_target_autotune=true (each OSD always has more than 12G).
>>> 
>>> For clients:
>>> client_cache_size=163840                                                    
>>>                                    
>>> client_oc_max_dirty=1048576000                                              
>>>                                      
>>> client_oc_max_dirty_age=50
>>> client_oc_max_objects=10000                                                 
>>>                                        
>>> client_oc_size=2097152000                                                   
>>>                                 
>>> client_oc_target_dirty=838860800
>>> 
>>>      Evaluate, following the documentation, which of these variables makes 
>>> sense for your cluster.
>>> 
>>>      For the backup scenario, I imagine that decreasing the size and 
>>> min_size values will change the impact. However, you must evaluate your 
>>> needs for these settings.
>>> 
>>> 
>>> Rafael.
>>> 
>>>  
>>> 
>>> De: "Kyriazis, George" <george.kyria...@intel.com>
>>> Enviada: 2024/10/02 13:06:09
>>> Para: ebl...@nde.ag, ceph-users@ceph.io
>>> Assunto: [ceph-users] Re: Question about speeding hdd based cluster
>>>  
>>> Thank you all.
>>> 
>>> The cluster is used mostly for backup of large files currently, but we are 
>>> hoping to use it for home directories (compiles, etc.) soon. Most usage 
>>> would be for large files, though.
>>> 
>>> What I've observed with its current usage is that ceph rebalances, and 
>>> proxmox-initiated VM backups bring the storage to its knees.
>>> 
>>> Would a safe approach be to move the metadata pool to ssd first, see how it 
>>> goes (since it would be cheaper), and then add DB/WAL disks? How would ceph 
>>> behave if we are adding DB/WAL disks "slowly" (ie one node at a time)? We 
>>> have about 100 OSDs (mix hdd/ssd) spread across about 25 hosts. Hosts are 
>>> server-grade with plenty of memory and processing power.
>>> 
>>> Thank you!
>>> 
>>> George
>>> 
>>> 
>>> > -----Original Message-----
>>> > From: Eugen Block <ebl...@nde.ag>
>>> > Sent: Wednesday, October 2, 2024 2:18 AM
>>> > To: ceph-users@ceph.io
>>> > Subject: [ceph-users] Re: Question about speeding hdd based cluster
>>> >
>>> > Hi George,
>>> >
>>> > the docs [0] strongly recommend to have dedicated SSD or NVMe OSDs for
>>> > the metadata pool. You'll also benefit from dedicated DB/WAL devices.
>>> > But as Joachim already stated, it depends on a couple of factors like the
>>> > number of clients, the load they produce, file sizes etc. There's no easy 
>>> > answer.
>>> >
>>> > Regards,
>>> > Eugen
>>> >
>>> > [0] https://docs.ceph.com/en/latest/cephfs/createfs/#creating-pools
>>> >
>>> > Zitat von Joachim Kraftmayer <joachim.kraftma...@clyso.com>:
>>> >
>>> > > Hi Kyriazis,
>>> > >
>>> > > depends on the workload.
>>> > > I would recommend to add ssd/nvme DB/WAL to each osd.
>>> > >
>>> > >
>>> > >
>>> > > Joachim Kraftmayer
>>> > >
>>> > > www.clyso.com <http://www.clyso.com/>
>>> > >
>>> > > Hohenzollernstr. 27, 80801 Munich
>>> > >
>>> > > Utting a. A. | HR: Augsburg | HRB: 25866 | USt. ID-Nr.: DE2754306
>>> > >
>>> > > Kyriazis, George <george.kyria...@intel.com> schrieb am Mi., 2. Okt.
>>> > > 2024,
>>> > > 07:37:
>>> > >
>>> > >> Hello ceph-users,
>>> > >>
>>> > >> I’ve been wondering…. I have a proxmox hdd-based cephfs pool with no
>>> > >> DB/WAL drives. I also have ssd drives in this setup used for other 
>>> > >> pools.
>>> > >>
>>> > >> What would increase the speed of the hdd-based cephfs more, and in
>>> > >> what usage scenarios:
>>> > >>
>>> > >> 1. Adding ssd/nvme DB/WAL drives for each node 2. Moving the metadata
>>> > >> pool for my cephfs to ssd 3. Increasing the performance of the
>>> > >> network. I currently have 10gbe links.
>>> > >>
>>> > >> It doesn’t look like the network is currently saturated, so I’m
>>> > >> thinking
>>> > >> (3) is not a solution. However, if I choose any of the other
>>> > >> options, would I need to also upgrade the network so that the network
>>> > >> does not become a bottleneck?
>>> > >>
>>> > >> Thank you!
>>> > >>
>>> > >> George
>>> > >>
>>> > >> _______________________________________________
>>> > >> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
>>> > >> email to ceph-users-le...@ceph.io
>>> > >>
>>> > > _______________________________________________
>>> > > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
>>> > > email to ceph-users-le...@ceph.io
>>> >
>>> >
>>> > _______________________________________________
>>> > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an 
>>> > email to
>>> > ceph-users-le...@ceph.io
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to 
>>> ceph-users-leave@ceph.io_______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io>
>>> To unsubscribe send an email to ceph-users-le...@ceph.io 
>>> <mailto:ceph-users-le...@ceph.io>

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to