nic (Which was trated as a
downgrade)
In Ubuntu maling lists they said those packages are "wrongly versioned"
I think the names should be 13.2.1-1ubuntu16.04-xenial and
13.2.1.ubuntu18.04-bionic.
--
Alfredo Daniel Rezinovsky
Director de Tecnologías de Información y Comunicaciones
Facultad
both in ubuntu 16.04 and 18.04 ceph-mgr fail to starts when package
python-routes is not installed
Some python packages are listed as dependencies for ceph-mgr but
python-routes is missing and must be installed manually for ceph-mgr to
work.
--
Alfredo Daniel Rezinovsky
Director de
On 20/08/18 06:44, John Spray wrote:
On Sun, Aug 19, 2018 at 9:21 PM Alfredo Daniel Rezinovsky
wrote:
both in ubuntu 16.04 and 18.04 ceph-mgr fail to starts when package
python-routes is not installed
I guess you mean that the dashboard doesn't work, as opposed to the
whole ceph-mgr pr
there and wont upgrade to beaver ones
because the names means downgrade.
- Original Message -
From: "Alfredo Daniel Rezinovsky"
To: "ceph-users"
Sent: Sunday, August 19, 2018 10:15:00 PM
Subject: [ceph-users] packages names for ubuntu/debian
Last packages for ubun
in an SDD partition.
--
Alfredo Daniel Rezinovsky
Director de Tecnologías de Información y Comunicaciones
Facultad de Ingeniería - Universidad Nacional de Cuyo
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/
15:18 systemd
-rw-r--r-- 1 ceph ceph 10 Aug 15 11:05 type
-rw-r--r-- 1 ceph ceph 2 Aug 15 11:05 whoami
Is a bluestore with block.db in a SSD. I'm not trusting the SSD and I
want to remove the block.db without destroying and creating the OSD.
There's a way to do this.
--
Alfredo D
ere)
Which kernel are you running?
Paul
2018-08-21 21:41 GMT+02:00 Alfredo Daniel Rezinovsky
:
Nope. I have plenty of RAM. 8Gb for 3 OSDs per node. Most of it used for
buffering.
On 21/08/18 16:09, Paul Emmerich wrote:
Are you running tight on memory?
Paul
2018-08-21 20:37 GMT+02:00 Alfr
rnal in an SDD partition.
--
Alfredo Daniel Rezinovsky
Director de Tecnologías de Información y Comunicaciones
Facultad de Ingeniería - Universidad Nacional de Cuyo
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/
d herein and
attached is confidential and the property of Beyond Hosting. Any
unauthorized copying, forwarding, printing, and/or disclosing
any information related to this email is prohibited. If you received
this message in error, please contact the sender and destroy all
copies of this email
and/or disclosing
any information related to this email is prohibited. If you received
this message in error, please contact the sender and destroy all
copies of this email and any attachment(s).
On Mon, Aug 27, 2018 at 10:36 PM Alfredo Daniel Rezinovsky
<mailto:alfred
], logical extent
0x1e000~1000, object #2:fd955b81:::1729cdb.0006
It happens sometimes, in all my OSDs.
Bluestore OSDs with data in HDD and block.db in SSD
After running pg repair the pgs were always repaired.
running ceph in ubuntu 13.2.1-1bionic
--
Alfredo Daniel Rezinovsky
Director de
Can anyone add me to this slack?
with my email alfrenov...@gmail.com
Thanks.
--
Alfredo Daniel Rezinovsky
Director de Tecnologías de Información y Comunicaciones
Facultad de Ingeniería - Universidad Nacional de Cuyo
___
ceph-users mailing list
ceph
Paul Emmerich wrote:
We built a work-around here: https://github.com/ceph/ceph/pull/23273
Which hasn't been backported, but we'll ship 13.2.2 in our Debian
packages for the croit OS image.
Paul
2018-09-18 21:10 GMT+02:00 Alfredo Daniel Rezinovsky
:
Changed all my hardware. Now I have
stribution kernel.
Paul
2018-09-18 21:23 GMT+02:00 Alfredo Daniel Rezinovsky
:
MOMENT !!!
"Some kernels (4.9+) sometime fail to return data when reading from a block
device under memory pressure."
I dind't knew that was the problem. Can't I just dowgrade the kernel?
Ther
Tried 4.17 with the same problem
Just downgraded to 4.8. Let's see if no more 0x67... appears
On 18/09/18 16:28, Alfredo Daniel Rezinovsky wrote:
I started with this after upgrade to bionic. I had Xenial with lts
kernels (4.13) without problem.
I will try to change to ubuntu 4.13 and
I have ubuntu servers.
With ukuu I installed kernel 4.8.17-040817 (The last < 4.9 available
kernel) and I haven't any 0x6706be76 crc since.
Nor any inconsistence.
On 19/09/18 12:01, Alfredo Daniel Rezinovsky wrote:
Tried 4.17 with the same problem
Just downgraded to 4.8. Let
Cluster with 4 nodes
node 1: 2 HDDs
node 2: 3 HDDs
node 3: 3 HDDs
node 4: 2 HDDs
After a problem with upgrade from 13.2.1 to 13.2.2 (I restarted the
nodes 1 at a time)
I upgraded with ubuntu apt-get upgrade. I had 1 acvive mds at a time
when did the upgrade.
All MDSs stopped working
Statu
Cluster with 4 nodes
node 1: 2 HDDs
node 2: 3 HDDs
node 3: 3 HDDs
node 4: 2 HDDs
After a problem with upgrade from 13.2.1 to 13.2.2 (I restarted the
nodes 1 at a time, think that was the problem)
I upgraded with ubuntu apt-get upgrade. I had 1 active mds at a time
when did the upgrade.
All
queue
disk format. Please downgrading mds to 13.2.1, then run 'ceph mds
repaired cephfs_name:0'.
Regards
Yan, Zheng
On Mon, Oct 8, 2018 at 9:20 AM Alfredo Daniel Rezinovsky
wrote:
Cluster with 4 nodes
node 1: 2 HDDs
node 2: 3 HDDs
node 3: 3 HDDs
node 4: 2 HDDs
After a problem with up
On 08/10/18 09:45, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel Rezinovsky
wrote:
On 08/10/18 07:06, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 5:43 PM Sergey Malinin wrote:
On 8.10.2018, at 12:37, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 4:37 PM Sergey Malinin wrote
On 08/10/18 10:32, Sergey Malinin wrote:
On 8.10.2018, at 16:07, Alfredo Daniel Rezinovsky
mailto:alfrenov...@gmail.com>> wrote:
So I can stopt cephfs-data-scan, run the import, downgrade, and then
reset the purge queue?
I suggest that you backup metadata pool so that in c
On 08/10/18 10:20, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 9:07 PM Alfredo Daniel Rezinovsky
wrote:
On 08/10/18 09:45, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel Rezinovsky
wrote:
On 08/10/18 07:06, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 5:43 PM Sergey Malinin
On 08/10/18 11:47, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 9:46 PM Alfredo Daniel Rezinovsky
wrote:
On 08/10/18 10:20, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 9:07 PM Alfredo Daniel Rezinovsky
wrote:
On 08/10/18 09:45, Yan, Zheng wrote:
On Mon, Oct 8, 2018 at 6:40 PM Alfredo Daniel
It seems my purge_queue journal is damaged. Even if I reset it keeps
damaged.
What means inotablev mismatch ?
2018-10-08 16:40:03.144 7f05b6099700 -1 log_channel(cluster) log [ERR] :
journal replay inotablev mismatch 1 -> 42160
/build/ceph-13.2.1/src/mds/journal.cc: In function 'void
EMetaBl
On 08/10/18 17:41, Sergey Malinin wrote:
On 8.10.2018, at 23:23, Alfredo Daniel Rezinovsky
mailto:alfrenov...@gmail.com>> wrote:
I need the data, even if it's read only.
After full data scan you should have been able to boot mds 13.2.2 and
mount the fs.
The problem start
.2.1 would be safe.
--
Patrick Donnelly
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-c
ces in use).
Hope this helps.
--
Alfredo Daniel Rezinovsky
Director de Tecnologías de Información y Comunicaciones
Facultad de Ingeniería - Universidad Nacional de Cuyo
<>___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/l
I've found only the official docs about bluestore compression. Seems simple.
I'ts stable enough to set it up in production servers?
Can be easilly disabled in case of trouble?
--
Alfrenovsky
___
ceph-users mailing list
ceph-users@lists.ceph.com
http:
I migrated my OSDs from filestore to bluestore.
Each node now has 1 SSD with the OS and the BlockDBs and 3 HDDs with
bluestore data.
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sdd 8:48 0 2.7T 0 disk
|-sdd2 8:50 0 2.7T 0 part
`-sdd1 8:49 0 100M 0 part /var/lib/c
On 13/06/18 01:03, Konstantin Shalygin wrote:
Each node now has 1 SSD with the OS and the BlockDBs and 3 HDDs with
bluestore data.
Very. Very bad idea. When your ssd/nvme dead you lost your linux box.
I have 3 boxes. And I'm installing a new one. Any box can be lost
without data problem.
amount of data; you could try
compacting the rocksdb instance etc. But if reading 800MB is
noticeable I would start wondering about the quality of your disks as
a journal or rocksdb device.
-Greg
On Tue, Jun 12, 2018 at 2:23 PM Alfredo Daniel Rezinovsky
<mailto:alfredo.rezi
On 18/06/18 09:09, Alfredo Deza wrote:
On Fri, Jun 15, 2018 at 11:59 AM, Alfredo Daniel Rezinovsky
wrote:
Too long is 120 seconds
The DB is in SSD devices. The devices are fast. The process OSD reads about
800Mb but I cannot be sure from where.
You didn't mention what version of Cep
autoscale-status reports some of my PG_NUMs are way too big
I have 256 and need 32
POOL SIZE TARGET SIZE RATE RAW CAPACITY RATIO TARGET
RATIO PG_NUM NEW PG_NUM AUTOSCALE
rbd 1214G 3.0 56490G
0.0645 256 32 warn
If
On 15/4/19 06:54, Jasper Spaans wrote:
On 14/04/2019 17:05, Alfredo Daniel Rezinovsky wrote:
autoscale-status reports some of my PG_NUMs are way too big
I have 256 and need 32
POOL SIZE TARGET SIZE RATE RAW CAPACITY RATIO TARGET
RATIO PG_NUM NEW PG_NUM AUTOSCALE
rbd
With 3 monitors, paxos needs at least 2 to reach consensus about the
cluster status
With 4 monitors, more than half is 3. The only problem I can see here is
that I will have only 1 spare monitor.
There's any other problem with and even number of monitors?
--
Alfrenovsky
I have a 4 node cluster. Each one has 1 SSD for the OS and block.dbs
(50Gb partition for each OSD), one 4Tb hdd and two 8Tb hdds.
I have 15% iowait average.
In any other server 15% seems too much. But ceph is a storage service
cluster.
There's a way to minimize the iowait or to better measur
36 matches
Mail list logo