[ceph-users] packages names for ubuntu/debian

2018-08-19 Thread Alfredo Daniel Rezinovsky
nic (Which was trated as a downgrade) In Ubuntu maling lists they said those packages are "wrongly versioned" I think the names should be 13.2.1-1ubuntu16.04-xenial and 13.2.1.ubuntu18.04-bionic. -- Alfredo Daniel Rezinovsky Director de Tecnologías de Información y Comunicaciones Facultad

[ceph-users] missing dependecy in ubuntu packages

2018-08-19 Thread Alfredo Daniel Rezinovsky
both in ubuntu 16.04 and 18.04 ceph-mgr fail to starts when package python-routes is not installed Some python packages are listed as dependencies for ceph-mgr but python-routes is missing and must be installed manually for ceph-mgr to work. -- Alfredo Daniel Rezinovsky Director de

Re: [ceph-users] missing dependecy in ubuntu packages

2018-08-20 Thread Alfredo Daniel Rezinovsky
On 20/08/18 06:44, John Spray wrote: On Sun, Aug 19, 2018 at 9:21 PM Alfredo Daniel Rezinovsky wrote: both in ubuntu 16.04 and 18.04 ceph-mgr fail to starts when package python-routes is not installed I guess you mean that the dashboard doesn't work, as opposed to the whole ceph-mgr pr

Re: [ceph-users] packages names for ubuntu/debian

2018-08-20 Thread Alfredo Daniel Rezinovsky
there and wont upgrade to beaver ones because the names means downgrade. - Original Message - From: "Alfredo Daniel Rezinovsky" To: "ceph-users" Sent: Sunday, August 19, 2018 10:15:00 PM Subject: [ceph-users] packages names for ubuntu/debian Last packages for ubun

[ceph-users] fixable inconsistencies but more appears

2018-08-21 Thread Alfredo Daniel Rezinovsky
in an SDD partition. -- Alfredo Daniel Rezinovsky Director de Tecnologías de Información y Comunicaciones Facultad de Ingeniería - Universidad Nacional de Cuyo ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/

[ceph-users] There's a way to remove the block.db ?

2018-08-21 Thread Alfredo Daniel Rezinovsky
15:18 systemd -rw-r--r-- 1 ceph ceph  10 Aug 15 11:05 type -rw-r--r-- 1 ceph ceph   2 Aug 15 11:05 whoami Is a bluestore with block.db in a SSD. I'm not trusting the SSD and I want to remove the block.db without destroying and creating the OSD. There's a way to do this. -- Alfredo D

Re: [ceph-users] fixable inconsistencies but more appears

2018-08-21 Thread Alfredo Daniel Rezinovsky
ere) Which kernel are you running? Paul 2018-08-21 21:41 GMT+02:00 Alfredo Daniel Rezinovsky : Nope. I have plenty of RAM. 8Gb for 3 OSDs per node. Most of it used for buffering. On 21/08/18 16:09, Paul Emmerich wrote: Are you running tight on memory? Paul 2018-08-21 20:37 GMT+02:00 Alfr

Re: [ceph-users] fixable inconsistencies but more appears

2018-08-27 Thread Alfredo Daniel Rezinovsky
rnal in an SDD partition. -- Alfredo Daniel Rezinovsky Director de Tecnologías de Información y Comunicaciones Facultad de Ingeniería - Universidad Nacional de Cuyo ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/

Re: [ceph-users] Bluestore crashing constantly with load on newly created cluster/host.

2018-08-27 Thread Alfredo Daniel Rezinovsky
d herein and attached is confidential and the property of Beyond Hosting. Any unauthorized copying, forwarding, printing, and/or disclosing any information related to this email is prohibited. If you received this message in error, please contact the sender and destroy all copies of this email

Re: [ceph-users] Bluestore crashing constantly with load on newly created cluster/host.

2018-08-27 Thread Alfredo Daniel Rezinovsky
and/or disclosing any information related to this email is prohibited. If you received this message in error, please contact the sender and destroy all copies of this email and any attachment(s). On Mon, Aug 27, 2018 at 10:36 PM Alfredo Daniel Rezinovsky <mailto:alfred

[ceph-users] No fix for 0x6706be76 CRCs ?

2018-09-18 Thread Alfredo Daniel Rezinovsky
], logical extent 0x1e000~1000, object #2:fd955b81:::1729cdb.0006 It happens sometimes, in all my OSDs. Bluestore OSDs with data in HDD and block.db in SSD After running pg repair the pgs were always repaired. running ceph in ubuntu 13.2.1-1bionic -- Alfredo Daniel Rezinovsky Director de

[ceph-users] https://ceph-storage.slack.com

2018-09-18 Thread Alfredo Daniel Rezinovsky
Can anyone add me to this slack? with my email alfrenov...@gmail.com Thanks. -- Alfredo Daniel Rezinovsky Director de Tecnologías de Información y Comunicaciones Facultad de Ingeniería - Universidad Nacional de Cuyo ___ ceph-users mailing list ceph

Re: [ceph-users] No fix for 0x6706be76 CRCs ?

2018-09-18 Thread Alfredo Daniel Rezinovsky
Paul Emmerich wrote: We built a work-around here: https://github.com/ceph/ceph/pull/23273 Which hasn't been backported, but we'll ship 13.2.2 in our Debian packages for the croit OS image. Paul 2018-09-18 21:10 GMT+02:00 Alfredo Daniel Rezinovsky : Changed all my hardware. Now I have

Re: [ceph-users] No fix for 0x6706be76 CRCs ?

2018-09-18 Thread Alfredo Daniel Rezinovsky
stribution kernel. Paul 2018-09-18 21:23 GMT+02:00 Alfredo Daniel Rezinovsky : MOMENT !!! "Some kernels (4.9+) sometime fail to return data when reading from a block device under memory pressure." I dind't knew that was the problem. Can't I just dowgrade the kernel? Ther

Re: [ceph-users] No fix for 0x6706be76 CRCs ?

2018-09-19 Thread Alfredo Daniel Rezinovsky
Tried 4.17 with the same problem Just downgraded to 4.8. Let's see if no more 0x67... appears On 18/09/18 16:28, Alfredo Daniel Rezinovsky wrote: I started with this after upgrade to bionic. I had Xenial with lts kernels (4.13) without problem. I will try to change to ubuntu 4.13 and

Re: [ceph-users] No fix for 0x6706be76 CRCs ? [SOLVED] (WORKAROUND)

2018-09-21 Thread Alfredo Daniel Rezinovsky
I have ubuntu servers. With ukuu I installed kernel 4.8.17-040817 (The last < 4.9 available kernel) and I haven't any 0x6706be76 crc since. Nor any inconsistence. On 19/09/18 12:01, Alfredo Daniel Rezinovsky wrote: Tried 4.17 with the same problem Just downgraded to 4.8. Let

[ceph-users] Error in MDS (laggy or creshed)

2018-10-07 Thread Alfredo Daniel Rezinovsky
Cluster with 4 nodes node 1: 2 HDDs node 2: 3 HDDs node 3: 3 HDDs node 4: 2 HDDs After a problem with upgrade from 13.2.1 to 13.2.2 (I restarted the nodes 1 at a time) I upgraded with ubuntu apt-get upgrade. I had 1 acvive mds at a time when did the upgrade. All MDSs stopped working Statu

[ceph-users] Error in MDS (laggy or creshed)

2018-10-07 Thread Alfredo Daniel Rezinovsky
Cluster with 4 nodes node 1: 2 HDDs node 2: 3 HDDs node 3: 3 HDDs node 4: 2 HDDs After a problem with upgrade from 13.2.1 to 13.2.2 (I restarted the nodes 1 at a time, think that was the problem) I upgraded with ubuntu apt-get upgrade. I had 1 active mds at a time when did the upgrade. All

Re: [ceph-users] Error in MDS (laggy or creshed)

2018-10-07 Thread Alfredo Daniel Rezinovsky
queue disk format. Please downgrading mds to 13.2.1, then run 'ceph mds repaired cephfs_name:0'. Regards Yan, Zheng On Mon, Oct 8, 2018 at 9:20 AM Alfredo Daniel Rezinovsky wrote: Cluster with 4 nodes node 1: 2 HDDs node 2: 3 HDDs node 3: 3 HDDs node 4: 2 HDDs After a problem with up

Re: [ceph-users] MDS damaged after mimic 13.2.1 to 13.2.2 upgrade

2018-10-08 Thread Alfredo Daniel Rezinovsky
On 08/10/18 09:45, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel Rezinovsky wrote: On 08/10/18 07:06, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 5:43 PM Sergey Malinin wrote: On 8.10.2018, at 12:37, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 4:37 PM Sergey Malinin wrote

Re: [ceph-users] MDS damaged after mimic 13.2.1 to 13.2.2 upgrade

2018-10-08 Thread Alfredo Daniel Rezinovsky
On 08/10/18 10:32, Sergey Malinin wrote: On 8.10.2018, at 16:07, Alfredo Daniel Rezinovsky mailto:alfrenov...@gmail.com>> wrote: So I can stopt  cephfs-data-scan, run the import, downgrade, and then reset the purge queue? I suggest that you backup metadata pool so that in c

Re: [ceph-users] MDS damaged after mimic 13.2.1 to 13.2.2 upgrade

2018-10-08 Thread Alfredo Daniel Rezinovsky
On 08/10/18 10:20, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 9:07 PM Alfredo Daniel Rezinovsky wrote: On 08/10/18 09:45, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel Rezinovsky wrote: On 08/10/18 07:06, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 5:43 PM Sergey Malinin

Re: [ceph-users] MDS damaged after mimic 13.2.1 to 13.2.2 upgrade

2018-10-08 Thread Alfredo Daniel Rezinovsky
On 08/10/18 11:47, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 9:46 PM Alfredo Daniel Rezinovsky wrote: On 08/10/18 10:20, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 9:07 PM Alfredo Daniel Rezinovsky wrote: On 08/10/18 09:45, Yan, Zheng wrote: On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel

[ceph-users] MDSs still core dumping

2018-10-08 Thread Alfredo Daniel Rezinovsky
It seems my purge_queue journal is damaged. Even if I reset it keeps damaged. What means inotablev mismatch ? 2018-10-08 16:40:03.144 7f05b6099700 -1 log_channel(cluster) log [ERR] : journal replay inotablev mismatch 1 -> 42160 /build/ceph-13.2.1/src/mds/journal.cc: In function 'void EMetaBl

Re: [ceph-users] MDSs still core dumping

2018-10-08 Thread Alfredo Daniel Rezinovsky
On 08/10/18 17:41, Sergey Malinin wrote: On 8.10.2018, at 23:23, Alfredo Daniel Rezinovsky mailto:alfrenov...@gmail.com>> wrote: I need the data, even if it's read only. After full data scan you should have been able to boot mds 13.2.2 and mount the fs. The problem start

Re: [ceph-users] Don't upgrade to 13.2.2 if you use cephfs

2018-10-17 Thread Alfredo Daniel Rezinovsky
.2.1 would be safe. -- Patrick Donnelly ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-c

[ceph-users] slow MDS requests [Solved]

2018-06-07 Thread Alfredo Daniel Rezinovsky
ces in use). Hope this helps. -- Alfredo Daniel Rezinovsky Director de Tecnologías de Información y Comunicaciones Facultad de Ingeniería - Universidad Nacional de Cuyo <>___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/l

[ceph-users] bluestore compression stability

2018-06-11 Thread Alfredo Daniel Rezinovsky
I've found only the official docs about bluestore compression. Seems simple. I'ts stable enough to set it up in production servers? Can be easilly disabled in case of trouble? -- Alfrenovsky ___ ceph-users mailing list ceph-users@lists.ceph.com http:

[ceph-users] OSDs too slow to start

2018-06-12 Thread Alfredo Daniel Rezinovsky
I migrated my OSDs from filestore to bluestore. Each node now has 1 SSD with the OS and the BlockDBs and 3 HDDs with bluestore data. # lsblk NAME   MAJ:MIN RM   SIZE RO TYPE MOUNTPOINT sdd  8:48   0   2.7T  0 disk |-sdd2   8:50   0   2.7T  0 part `-sdd1   8:49   0   100M  0 part /var/lib/c

Re: [ceph-users] OSDs too slow to start

2018-06-13 Thread Alfredo Daniel Rezinovsky
On 13/06/18 01:03, Konstantin Shalygin wrote: Each node now has 1 SSD with the OS and the BlockDBs and 3 HDDs with bluestore data. Very. Very bad idea. When your ssd/nvme dead you lost your linux box. I have 3 boxes. And I'm installing a new one. Any box can be lost without data problem.

Re: [ceph-users] OSDs too slow to start

2018-06-15 Thread Alfredo Daniel Rezinovsky
amount of data; you could try compacting the rocksdb instance etc. But if reading 800MB is noticeable I would start wondering about the quality of your disks as a journal or rocksdb device. -Greg On Tue, Jun 12, 2018 at 2:23 PM Alfredo Daniel Rezinovsky <mailto:alfredo.rezi

Re: [ceph-users] OSDs too slow to start

2018-06-18 Thread Alfredo Daniel Rezinovsky
On 18/06/18 09:09, Alfredo Deza wrote: On Fri, Jun 15, 2018 at 11:59 AM, Alfredo Daniel Rezinovsky wrote: Too long is 120 seconds The DB is in SSD devices. The devices are fast. The process OSD reads about 800Mb but I cannot be sure from where. You didn't mention what version of Cep

[ceph-users] Decreasing pg_num

2019-04-14 Thread Alfredo Daniel Rezinovsky
autoscale-status reports some of my PG_NUMs are way too big I have 256 and need 32 POOL   SIZE  TARGET SIZE  RATE  RAW CAPACITY RATIO  TARGET RATIO  PG_NUM  NEW PG_NUM  AUTOSCALE rbd   1214G    3.0 56490G  0.0645   256  32  warn If

Re: [ceph-users] Decreasing pg_num

2019-04-15 Thread Alfredo Daniel Rezinovsky
On 15/4/19 06:54, Jasper Spaans wrote: On 14/04/2019 17:05, Alfredo Daniel Rezinovsky wrote: autoscale-status reports some of my PG_NUMs are way too big I have 256 and need 32 POOL   SIZE  TARGET SIZE  RATE  RAW CAPACITY RATIO  TARGET RATIO  PG_NUM  NEW PG_NUM  AUTOSCALE rbd

[ceph-users] even number of monitors

2019-08-05 Thread Alfredo Daniel Rezinovsky
With 3 monitors, paxos needs at least 2 to reach consensus about the cluster status With 4 monitors, more than half is 3. The only problem I can see here is that I will have only 1 spare monitor. There's any other problem with and even number of monitors? -- Alfrenovsky

[ceph-users] How much iowait is too much iowait?

2019-08-20 Thread Alfredo Daniel Rezinovsky
I have a 4 node cluster. Each one has 1 SSD for the OS and block.dbs (50Gb partition for each OSD), one 4Tb hdd and two 8Tb hdds. I have 15% iowait average. In any other server 15% seems too much. But ceph is a storage service cluster. There's a way to minimize the iowait or to better measur