[ceph-users] Get Instant HELP From Epson support For Epson Error Code 0x97.

2020-08-21 Thread mary smith
The Epson Error Code 0x97 is a kind of caution at whatever point you experience an inside equipment issue with your printer. It very well may be a result of a motherboard glitch or some other sort of inner segment breakdown. You should choose the guide from Epson support , straightforwardly from

[ceph-users] Re: pg stuck in unknown state

2020-08-21 Thread Michael Thomas
On 8/11/20 8:35 AM, Michael Thomas wrote: On 8/11/20 2:52 AM, Wido den Hollander wrote: On 11/08/2020 00:40, Michael Thomas wrote: On my relatively new Octopus cluster, I have one PG that has been perpetually stuck in the 'unknown' state.  It appears to belong to the device_health_metrics po

[ceph-users] does ceph RBD have the ability to load balance?

2020-08-21 Thread ??????????
Hello Cephers, I am a newcomer to Ceph.I know that RBD is a distributed block storage device and libvirt supports rbd pool type. Suppose I have a ceph cluster and use it to build a rbd pool to store virtual machine images. When multiple qemu-kvm clients use the image in the rbd pool to start

[ceph-users] Ceph raw capacity usage does not meet real pool storage usage

2020-08-21 Thread Davood Ghatreh
Hello everybody, I have a Ceph cluster (14.2.9) running with 18 X 600GB HDD OSDs. There are three pools (size:3, pg_num:64) with an image size of 200GB on each, and there are 6 servers connected to these images via iSCSI and storing about 20 VMs on them. Here is the output of "ceph df": POOLS: P

[ceph-users] Re: Adding OSD

2020-08-21 Thread jcharles
here yiu are The osd that have been added is osd.40 ssd, and it's an nautilus cluster Thanks for helping - Crush Map ID CLASS WEIGHT(compat) TYPE NAME -2633.37372 root HDD10 -33 6.67474 6.67474 host HDD10-ceph01 1 hdd10 1.66869 1.66869

[ceph-users] Re: Adding OSD

2020-08-21 Thread Nathan Fish
Please post the crush rules of your pools, and "ceph status" On Fri., Aug. 21, 2020, 3:33 p.m. , wrote: > Hello, > > I just add one new osd in one of my osd hosts. Crush map is updated but > nothing happen, the new osd is seen by the cluster, but norebalance occurs. > I have to admit I am a bit

[ceph-users] Adding OSD

2020-08-21 Thread jcharles
Hello, I just add one new osd in one of my osd hosts. Crush map is updated but nothing happen, the new osd is seen by the cluster, but norebalance occurs. I have to admit I am a bit stressed. I have my current OSD at 70-75% of capacity. May be should I add some osd to other hosts for the cluster

[ceph-users] [doc] drivegroups advanced case

2020-08-21 Thread Tony Liu
Hi, Regarding to the YAML in this section, https://ceph.readthedocs.io/en/latest/cephadm/drivegroups/#the-advanced-case Is the "rotational" supposed to be "1" meaning spinning HDD? Thanks! Tony ___ ceph-users mailing list -- ceph-users@ceph.io To unsu

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Eugen Block
Yes, I don’t see any reason why it wouldn’t be. The client performance can be controlled by max-backfills and recovery-max-active settings, there was a thread this week about those. Zitat von Matt Dunavant : Gotcha, 1 more question: During the process, data will be available right? Just pe

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Matt Dunavant
Gotcha, 1 more question: During the process, data will be available right? Just performance will be impacted by the rebalancing correct? ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Eugen Block
Although I’m confused about the error from crushtool it seems the results are fine. The show-mappings displays a long list of possible mappings to OSDs. If you provide num-rep 3 you should see three different OSDs in each line, and if the rule works correctly those OSDs never map to the sam

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Matt Dunavant
My replica size on the pool is 3, so I'll use that to test. There is no other type in my map like dc, rack, etc; just servers. Do you know what a successful run of the test command looks like? I just ran it myself and it spits out a number of crush rules (in this case 1024) and then ends with:

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Eugen Block
Thanks for the reply! I've pasted what I believe are the applicable parts of the crush map below. I see that the rule id is 0, but what is num-rep? num-rep is the number of replicas you want to test, so basically the size paramater of the pool this rule applies to. Do you have any hierach

[ceph-users] Re: OSDs get full with bluestore logs

2020-08-21 Thread Igor Fedotov
Indeed this assertion "FAILED assert(0 == "bluefs enospc")" indicates lack of free space for RocksDB at both main and DB volumes. OSD (RocksDB specifically) attempts to recover (and hence flush) some data on OSD restart and is unable to allocate space for that. Hence it crashes... What volu

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Matt Dunavant
Thanks for the reply! I've pasted what I believe are the applicable parts of the crush map below. I see that the rule id is 0, but what is num-rep? # begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1

[ceph-users] Re: BlueFS spillover detected, why, what?

2020-08-21 Thread Igor Fedotov
Can't say nothing about "write_buffer_size" tuning.. Never tried that. But I presume that these are *"max_bytes_for_level_base*" and *"**max_bytes_for_level_multiplier*" params which rather should be tuned to modify RocksDB level granularity. But I have no ideas how safe this is in a producti

[ceph-users] Re: Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Eugen Block
Hi, 1) I believe the correct way to fix this is by following the 5 step method in the documentation; Get, Decompile, Edit, Recompile, Set. Is that correct and is the line I should change 'choose_firstn' to 'chooseleaf_firstn'? Do I only make this change on 1 mon and it will propagate it t

[ceph-users] Editing Crush Map to fix osd_crush_chooseleaf_type = 0

2020-08-21 Thread Matt Dunavant
Hi all, We have a 12 OSD node cluster in which I just recently found out that 'osd_crush_chooseleaf_type = 0' made it's way into our ceph.conf file, probably from previous testing. I believe this is the reason a recent maintenance on an OSD node caused data to stop flowing. In researching how

[ceph-users] How do I configure a Roadrunner email account?

2020-08-21 Thread smith8395john
I am a professional person and dealing with small and big customers. So I need an emailing application to communicate with them online. On a daily basis, I have to make many business deals with the customers. So I need to send business deals details. So I have decided to set up a roadrunner emai

[ceph-users] Re: Remove Error - "Possible data damage: 2 pgs recovery_unfound"

2020-08-21 Thread Philipp Hocke
Hi, you could try following the troubleshooting pg section https://docs.ceph.com/docs/mimic/rados/troubleshooting/troubleshooting-pg/#unfound-objects I had some unfound objects a while ago and managed to restore them to a previous version. If you just want to "get rid" of the error and _really r

[ceph-users] Re: Remove Error - "Possible data damage: 2 pgs recovery_unfound"

2020-08-21 Thread Jonathan Sélea
Hi everyone, I just wanted to ask again about your oppinion about this "problem" that I have. Thankful for any answer! On 2020-08-19 13:39, Jonathan Sélea wrote: Good afternoon! I have a small Ceph-cluster running with Proxmox, and after an update on one of the nodes and a reboot. So far so

[ceph-users] Re: 5 pgs inactive, 5 pgs incomplete

2020-08-21 Thread Martin Palma
Sorry for stressing but it would help us a lot if someone with deeper knowledge could tell us if marking the PG on the secondary OSD will not render the whole CephFS pool unusable? We are aware that it could mean that some files will be lost or inconsistent but it will not affect all data in the po