Hello,
How to view latest OSD epoch value in luminous. Normally this can be found
by below command.
#ceph -s | grep osd
or
#ceph osd stat
Need to know how to find this from v12.1.0
Thanks
Jayaram
On Sun, Jul 2, 2017 at 6:11 PM, Marc Roos wrote:
>
> I have updated a test cluster by just updat
Hello,
Recently we got an underlying issue with osd.10 which mapped to /dev/sde .
So we tried to removed it from the crush
===
#systemctl stop ceph-osd@10.service
#for x in {10..10}; do ceph osd out $x;ceph osd crush remove osd.$x;ceph
auth del osd.$x;ceph osd rm osd.$x ;done
marked out osd.10.
Hello,
Env:- RHEL 7.2 , 3.10.0-327.el7.x86_64 , EC 4+1 , bluestore
We are writing to ceph via librados C API . Testing with rados no issues.
The same we tested with Jewel/kraken without any issue. Need your view how
to debug this issue?
>>
OSD.log
==
~~~
2017-09-18 14:51:59.895746 7f1e7
wrote:
> On 17-09-20 08:06, nokia ceph wrote:
>
> Hello,
>
> Env:- RHEL 7.2 , 3.10.0-327.el7.x86_64 , EC 4+1 , bluestore
>
> We are writing to ceph via librados C API . Testing with rados no issues.
>
>
> The same we tested with Jewel/kraken without any issue. Need
Hello,
I'm following
http://docs.ceph.com/docs/master/ceph-volume/lvm/prepare/#ceph-volume-lvm-prepare-bluestore
to create new OSD's.
I took the latest branch from https://shaman.ceph.com/repos/ceph/luminous/
# ceph -v
ceph version 12.2.1-851-g6d9f216
What I did, formatted the device.
#sgdisk
THanks brad, that got worked.. :)
On Fri, Dec 1, 2017 at 12:18 PM, Brad Hubbard wrote:
>
>
> On Thu, Nov 30, 2017 at 5:30 PM, nokia ceph
> wrote:
> > Hello,
> >
> > I'm following
> > http://docs.ceph.com/docs/master/ceph-volume/lvm/
> prepare/#cep
Hello Team,
I having a 5 node cluster running with kraken 11.2.0 EC 4+1.
My plan is to upgrade all 5 nodes to 12.2.2 Luminous without any downtime.
I tried on first node, below procedure.
commented below directive from ceph.conf
enable experimental unrecoverable data corrupting features = bluest
OSDs should come back up.
>
> On Fri, Dec 8, 2017 at 10:51 AM nokia ceph
> wrote:
>
>> Hello Team,
>>
>> I having a 5 node cluster running with kraken 11.2.0 EC 4+1.
>>
>> My plan is to upgrade all 5 nodes to 12.2.2 Luminous without any
>> downtim
Hello Team,
We aware that ceph-disk which is deprecated in 12.2.2 . As part of my
testing, I can still using this ceph-disk utility for creating OSD's in
12.2.2
Here I'm getting activation error on the second hit onwards.
First occurance OSD's creating without any issue.
===
Created tracker for this issue -- > http://tracker.ceph.com/issues/22354
Thanks
Jayaram
On Fri, Dec 8, 2017 at 9:49 PM, nokia ceph wrote:
> Hello Team,
>
> We aware that ceph-disk which is deprecated in 12.2.2 . As part of my
> testing, I can still using this ceph-disk utili
Hello,
i have installed Luminous 12.2.2 on a 5 node cluster with logical volume
OSDs.
I am trying to stop and start ceph on one of the nodes using systemctl
commands.
*systemctl stop ceph.target; systemctl start ceph.target*
When i stop ceph, all OSDs are stopped on the node properly.
But when i
rwx 1 root root 41 Jan 23 09:36 ceph-osd@12.service ->
/usr/lib/systemd/system/ceph-osd@.service
.
.
.
On Mon, Jan 8, 2018 at 3:49 PM, nokia ceph wrote:
> Hello,
>
> i have installed Luminous 12.2.2 on a 5 node cluster with logical volume
> OSDs.
> I am trying to stop and
Hi Team,
We need a mechanism to have some data cache on OSD build on bluestore . Is
there an option available to enable data cache?
With default configurations , OSD logs state that data cache is disabled by
default,
bluestore(/var/lib/ceph/osd/ceph-66) _set_cache_sizes cache_size 1073741824
Hi All,
We have 5 node clusters with EC 4+1 and use bluestore since last year from
Kraken.
Recently we migrated all our platforms to luminous 12.2.2 and finally all
OSDs migrated to ceph-volume simple type and on few platforms installed
ceph using ceph-volume .
Now we see two times more traffic i
performance. During rocksdb compaction the situation is worse.
Meanwhile we are building another platform creating osd using ceph-disk and
analyse on this.
Thanks,
Muthu
On Tuesday, February 20, 2018, Alfredo Deza wrote:
>
>
> On Mon, Feb 19, 2018 at 2:01 PM, nokia ceph
> wrote:
. If we consider
only lvm based system does this high iops because of dm-cache created for
each osd?.
Meanwhile i will update some graphs to show this once i have.
Thanks,
Muthu
On Tuesday, February 20, 2018, Alfredo Deza wrote:
>
>
> On Mon, Feb 19, 2018 at 9:29 PM, nokia ceph
> wr
.
Thanks,
Muthu
On Wed, Feb 21, 2018 at 6:57 PM, Alfredo Deza wrote:
>
>
> On Tue, Feb 20, 2018 at 9:33 PM, nokia ceph
> wrote:
>
>> Hi Alfredo Deza,
>>
>> I understand the point between lvm and simple however we see issue , was
>> it issue in lumin
Hi All,
We are having 5 node cluster with EC 4+1 . Each node has 68 HDD . Now we
are trying to add new node with 68 disks to the cluster .
We tried to add new node and created all OSDs in one go , the cluster
stopped all client traffic and does only backfilling .
Any procedure to add the new nod
:
> You can add new OSDs with 0 weight and edit below script to increase
> the osd weights instead of decreasing.
>
> https://github.com/cernceph/ceph-scripts/blob/master/
> tools/ceph-gentle-reweight
>
>
> On Wed, Apr 18, 2018 at 2:16 PM, nokia ceph
> wrote:
> > Hi
32 PM, Serkan Çoban
> wrote:
> >>
> >> You can add new OSDs with 0 weight and edit below script to increase
> >> the osd weights instead of decreasing.
> >>
> >> https://github.com/cernceph/ceph-scripts/blob/master/
> tools/ceph-gentle-reweight
Hi Team,
I was trying to modify the max_misplaced parameter in 12.2.4 as per
documentation , however not able to modify it with following error,
#ceph config set mgr mgr/balancer/max_misplaced .06
Invalid command: unused arguments: [u'.06']
config set : Set a configuration option at runtime (
active+remapped+backfilling
Thanks,
Muthu
On Fri, Apr 27, 2018 at 7:54 PM, John Spray wrote:
> On Fri, Apr 27, 2018 at 7:03 AM, nokia ceph
> wrote:
> > Hi Team,
> >
> > I was trying to modify the max_misplaced parameter in 12.2.4 as per
> > documentation , howe
Hi Team,
I was trying to forcefully lost the unfound objects using the below
commands mentioned in the documentation , it is not working in the latest
release , any prerequisites required for EC pool.
cn1.chn6m1c1ru1c1.cdn ~# *ceph pg 4.1206 mark_unfound_lost revert|delete*
-bash: delete: comman
Thank you , it works
On Tue, May 8, 2018 at 2:05 PM, Paul Emmerich
wrote:
> EC pools only support deleting unfound objects as there aren't multiple
> copies around that could be reverted to.
>
> ceph pg mark_unfound_lost delete
>
>
> Paul
>
> 2018-05-08 9:26 GMT
Hi Ceph users,
We have a cluster with 5 node (67 disks) and EC 4+1 configuration and
min_size set as 4.
Ceph version : 12.2.5
While executing one of our resilience usecase , making private interface
down on one of the node, till kraken we saw less outage in rados (60s) .
Now with luminous, we cou
stops or the network comes back up.
> There might be a seeing for how long an odd will try telling the mons it's
> up, but this isn't really a situation I've come across after initial
> testing and installation of nodes.
>
> On Tue, May 22, 2018, 1:47 AM nokia ceph wro
ioned and discussed multiple times on the ML.
>
>
> On Wed, May 23, 2018, 3:39 AM nokia ceph wrote:
>
>> Hi David Turner,
>>
>> This is our ceph config under mon section , we have EC 4+1 and set the
>> failure domain as host and osd_min_down_reporters to 4 (
please suggest other options which we can try.
thanks,
Muthu
On Wed, May 23, 2018 at 4:51 PM, nokia ceph
wrote:
> yes it is 68 disks , and will this mon_osd_reporter_subtree_level = host
> have any impact on mon_osd_ min_down_reporters ?
>
> And related to min_size , yes th
Hello,
On a fresh installation ceph kraken 11.2.0 , we are facing below error in
the "ceph -s" output.
0 -- 10.50.62.152:0/675868622 >> 10.50.62.152:6866/13884
conn(0x7f576c002750 :-1 s=STATE_CONNECTING_WAIT_BANNER_AND_IDENTIFY pgs=0
cs=0 l=1)._process_connection connect claims to be 10.50.6
have any other suggestion to how to skip this warning?
Thanks
On Mon, Feb 27, 2017 at 8:47 PM, Gregory Farnum wrote:
> On Sun, Feb 26, 2017 at 10:41 PM, nokia ceph
> wrote:
> > Hello,
> >
> > On a fresh installation ceph kraken 11.2.0 , we are facing below error in
Hello,
Env:- v11.2.0 - bluestore - EC 3 + 1
We are getting below entries both in /var/log/messages and osd logs. May I
know what is the impact of the below message and as these message were
flooded in osd and sys logs.
~~~
2017-03-01 13:00:59.938839 7f6c96915700 -1
bdev(/var/lib/ceph/osd/ceph-0
Hello,
Can we get any update for this problem?
Thanks
On Thu, Mar 2, 2017 at 2:16 PM, nokia ceph wrote:
> Hello,
>
> Env:- v11.2.0 - bluestore - EC 3 + 1
>
> We are getting below entries both in /var/log/messages and osd logs. May I
> know what is the impact of the below me
it retries 11*
2017-03-14 20:13:04.291160 7fee05294700 4 rocksdb: reusing log 85 from
recycle list
2017-03-14 20:13:04.291254 7fee05294700 4 rocksdb: [default] New memtable
created with log file: #89. Immutable memtables: 0.
=
Thanks
On Wed, Mar 15, 2017 at 11:18 AM, nokia ceph
wrote:
&
On Wed, 15 Mar 2017, Brad Hubbard wrote:
> > +ceph-devel
> >
> > On Wed, Mar 15, 2017 at 5:25 PM, nokia ceph
> wrote:
> > > Hello,
> > >
> > > We suspect these messages not only at the time of OSD creation. But in
> idle
> > > condit
in how this calculation happens?
>
> What calculation are you referring to?
>
> > Thanks
> >
> >
> > On Wed, Mar 15, 2017 at 7:54 PM, Sage Weil wrote:
> >>
> >> On Wed, 15 Mar 2017, Brad Hubbard wrote:
> >> > +ceph-devel
> >>
Sounds good :), Brad many thanks for the explanation .
On Thu, Mar 16, 2017 at 12:42 PM, Brad Hubbard wrote:
> On Thu, Mar 16, 2017 at 4:33 PM, nokia ceph
> wrote:
> > Hello Brad,
> >
> > I meant for this parameter bdev_aio_max_queue_depth , Sage suggested try
> >
Hello,
I made some changes in the below file on ceph kraken v11.2.0 source code as
per this article
https://github.com/ceph/ceph-ci/commit/wip-prune-past-intervals-kraken
..src/osd/PG.cc
..src/osd/PG.h
Is there any way to find which rpm got affected by these two files. I
believe it should be ce
s
> $ git log --grep="_simplify_past_intervals"
> $
>
> Given this commit is not in the ceph repo I would suggest we have never
> shipped
> an official rpm that contains this commit.
>
> It's not totally clear to me exactly what you are trying to achieve, ma
Thanks
On Thu, Mar 23, 2017 at 6:18 PM, Piotr Dałek
wrote:
> On 03/23/2017 01:41 PM, nokia ceph wrote:
>
>> Hey brad,
>>
>> Thanks for the info.
>>
>> Yea we know that these are test rpm's.
>>
>> The idea behind my question is if I made any chang
ote:
> On 03/23/2017 02:02 PM, nokia ceph wrote:
>
> Hello Piotr,
>>
>> We do customizing ceph code for our testing purpose. It's a part of our
>> R&D :)
>>
>> Recompiling source code will create 38 rpm's out of these I need to find
>> whi
t 5:41 AM, Brad Hubbard wrote:
> Oh wow, I completely misunderstood your question.
>
> Yes, src/osd/PG.cc and src/osd/PG.h are compiled into the ceph-osd binary
> which
> is included in the ceph-osd rpm as you said in your OP.
>
> On Fri, Mar 24, 2017 at 3:10 AM, nokia ceph
> wrote:
Piotr, thanks for the info.
Yea this method is time saving, but we are not started testing with build
from source method. We will consider this for our next part of testing :)
On Fri, Mar 24, 2017 at 1:17 PM, Piotr Dałek
wrote:
> On 03/23/2017 06:10 PM, nokia ceph wrote:
>
>>
Hello,
We are facing some performance issue with rados bench marking on a 5 node
cluster with PG num 4096 vs 8192.
As per the PG calculation below is our specification
Size OSD % Data Targets PG count
5 340 100 100 8192
5 340 100 50 4096
With 8192 PG count we got good performance with 409
ompiled like this.
Thanks
On Mon, Mar 27, 2017 at 5:04 AM, Brad Hubbard wrote:
>
>
> On Fri, Mar 24, 2017 at 6:49 PM, nokia ceph
> wrote:
> > Brad, cool now we are on the same track :)
> >
> > So whatever change we made after this location src/* as it mapped to
Hello,
We manually fixed the issue and below is our analysis.
Due to high CPU utilisation we stopped the ceph-mgr on all our cluster.
On one of our cluster we saw high memory usage by OSDs some grater than 5GB
causing OOM , resulting in process kill.
The memory was released immediately when the
Hello,
Env:-
5 node, EC 4+1 bluestore kraken v11.2.0 , RHEL7.2
As part of our resillency testing with kraken bluestore, we face more PG's
were in incomplete+remapped state. We tried to repair each PG using "ceph
pg repair " still no luck. Then we planned to remove incomplete PG's
using below proc
rocedure to safely remove an affected PG? As per my
understanding I'm aware about this command.
===
#ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph --pgid 1.e4b --op
remove
===
Awaiting for your suggestions to proceed.
Thanks
On Thu, Mar 30, 2017 at 7:32 AM, Brad Hubbard wrote
Hello,
We are getting below trace on failed OSD's . Can you please explain from
the below code why this issue happening. We suspect it could be because of
underlying HW issue. We can't find anything from the syslogs. All the OSD
disk are in healthy condition.
Link :- https://fossies.org/linux/cep
g 1.93f mark_unfound_lost delete { data loss }
Need your views on this, to how to clear the unfound issues without data
loss.
Thanks
Jayaram
On Mon, Apr 3, 2017 at 6:50 PM, Sage Weil wrote:
> On Fri, 31 Mar 2017, nokia ceph wrote:
> > Hello Brad,
> > Many thanks of the info
Hello,
Env:- 11.2.0
bluestore, EC 4+1 , RHEL7.2
We are facing one OSD's booting again and again which caused the cluster
crazy :( . As you can see one PG got in inconsistent state while we tried
to repair that partular PG, as its primary OSD's went down. After some time
we found some tr
Hello,
Env;- Bluestore EC 4+1 v11.2.0 RHEL7.3 16383 PG
We did our resiliency testing and found OSD's keeps on flapping and
cluster went to error state.
What we did:-
1. we have 5 node cluster
2. poweroff/stop ceph.target on last node and waited everything seems to
reach back to normal.
3. Th
Hello Mark,
I can able to reproduce this problem everytime.
Env:-- 5 node, v12.0.3, EC 4+1 bluestore , RHEL 7.3 - 3.10.0-514.el7.x86_64
Tested with debug bluestore = 20...
>From ceph watch
===
2017-05-30 08:57:33.510794 mon.0 [INF] pgmap v15649: 8192 pgs: 8192
active+clean; 774 GB data
e ,
#gcore
or using wallclock profiler, I'm not much aware how to use this tool.
Thanks
Jayaram
On Tue, May 30, 2017 at 6:57 PM, Mark Nelson wrote:
> On 05/30/2017 05:07 AM, nokia ceph wrote:
>
>> Hello Mark,
>>
>> I can able to reproduce this problem everytime.
&
9)
Thanks
Jayaram
On Tue, May 30, 2017 at 7:33 PM, nokia ceph
wrote:
> Hello Mark,
>
> Yes this issue happens once the test/write started after 60 secs which
> correspond config value -- "threadpool_default_timeout = 60 " . Do you
> require the down OSD coredump to
Hello Mark,
Raised tracker for the issue -- http://tracker.ceph.com/issues/20222
Jake can you share the restart_OSD_and_log-this.sh script
Thanks
Jayaram
On Wed, Jun 7, 2017 at 9:40 PM, Jake Grimmett wrote:
> Hi Mark & List,
>
> Unfortunately, even when using yesterdays master version of cep
ce the problem would be much appreciated!
>
> Mark
>
> On 06/08/2017 06:08 AM, nokia ceph wrote:
>
>> Hello Mark,
>>
>> Raised tracker for the issue -- http://tracker.ceph.com/issues/20222
>>
>> Jake can you share the restart_OSD_and_log-this.sh
OSD >> /root/osd_restart_log
> echo "OSD" $OSD "is down, restarting.."
> OSDHOST=`ceph osd find $OSD | grep host | awk -F '"' '{print $4}'`
> ssh $OSDHOST systemctl restart ceph-osd@$OSD
> sleep 30
>
Hello,
Some osd's not getting activated after a reboot operation which cause that
particular osd's landing in failed state.
Here you can see mount points were not getting updated to osd-num and
mounted as a incorrect mount point, which caused osd. can't able to
mount/activate the osd's.
Env:- RH
can test by
> chowning the journal block device and try to start the OSD again.
>
> Alternatively if you want to see more information, you can start the
> daemon manually as opposed to starting it through systemd and see what its
> output looks like.
>
> On Tue, Jun 13, 2017 at 6
Hi,
We have a 5 node EC 4+1 cluster with 335 OSDs running Kraken Bluestore
11.2.0.
There was a disk failure on one of the OSDs and the disk was replaced.
After which it was noticed that there was a ~30TB drop in the MAX_AVAIL
value for the pool storage details on output of 'ceph df'
Even though th
Hi Team,
After upgrading from Luminous to Nautilus , we see 654 pgs not
deep-scrubbed in time error in ceph status . How can we disable this flag?
. In our setup we disable deep-scrubbing for performance issues.
Thanks,
Muthu
___
ceph-users mailing list
your osd.
>
>
>
> For disable deep-scrub you can use “ceph osd set nodeep-scrub” , Also you
> can setup deep-scrub with threshold .
>
> #Start Scrub 22:00
>
> osd scrub begin hour = 22
>
> #Stop Scrub 8
>
> osd scrub end hour = 8
>
> #Scrub Load 0.5
>
&g
Hi Team,
We have a requirement to create multiple copies of an object and currently
we are handling it in client side to write as separate objects and this
causes huge network traffic between client and cluster.
Is there possibility of cloning an object to multiple copies using librados
api?
Pleas
will clone/copy multiple objects and stores inside the cluster.
Thanks,
Muthu
On Fri, Jun 28, 2019 at 9:23 AM Brad Hubbard wrote:
> On Thu, Jun 27, 2019 at 8:58 PM nokia ceph
> wrote:
> >
> > Hi Team,
> >
> > We have a requirement to create multiple copies of an
his by default. For each replicated pool, you can set
> the 'size' which is the number of copies you want Ceph to maintain. The
> accepted norm for replicas is 3, but you can set it higher if you want to
> incur the performance penalty.
>
> On Mon, Jul 1, 2019, 6:01 AM nok
ffic as the machine running the object class will still need to
> > > connect to the relevant primary osd and send the write (presumably in
> > > some situations though this will be the same machine).
> > >
> > > On Tue, Jul 2, 2019 at 4:08 PM nokia ceph
&g
ese are the fadvise flags we have in various
> places that let you specify things like not to cache the data.
> Probably leave them unset.
>
> -Greg
>
>
>
> On Wed, Jul 3, 2019 at 2:47 AM nokia ceph
> wrote:
> >
> > Hi Greg,
> >
> > Can you please sha
Hi Team,
After upgrading our cluster from 14.2.1 to 14.2.2 , the cluster moved to
warning state with following error
cn1.chn6m1c1ru1c1.cdn ~# ceph status
cluster:
id: e9afb5f3-4acf-421a-8ae6-caaf328ef888
health: HEALTH_WARN
Legacy BlueStore stats reporting detected on 33
chen
> www.croit.io
> Tel: +49 89 1896585 90
>
>
> On Fri, Jul 19, 2019 at 1:35 PM nokia ceph
> wrote:
>
>> Hi Team,
>>
>> After upgrading our cluster from 14.2.1 to 14.2.2 , the cluster moved to
>> warning state with following error
>>
>> cn
:
> bluestore warn on legacy statfs = false
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io
> Tel: +49 89 1896585 90
>
>
> On Fri, Jul 19, 2019
Hi Greg,
We were trying to implement this however having issues in assigning the
destination object name with this api.
There is a rados command "rados -p cp " , is
there any librados api equivalent to this ?
Thanks,
Muthu
On Fri, Jul 5, 2019 at 4:00 PM nokia ceph wrote:
> T
use librados.h in
our client to communicate with ceph cluster.
Also any equivalent librados api for the command rados -p poolname
Thanks,
Muthu
On Wed, Jul 31, 2019 at 11:13 PM Gregory Farnum wrote:
>
>
> On Wed, Jul 31, 2019 at 1:32 AM nokia ceph
> wrote:
>
>> Hi Greg,
&g
Hi Team,
Could you please help us in understanding the write iops inside ceph
cluster . There seems to be mismatch in iops between theoretical and what
we see in disk status.
Our platform 5 node cluster 120 OSDs, with each node having 24 disks HDD (
data, rcoksdb and rocksdb.WAL all resides in th
Thank you Greg, it is now clear for us and the option is only available in
C++ , we need to rewrite the client code with c++ .
Thanks,
Muthu
On Fri, Aug 2, 2019 at 1:05 AM Gregory Farnum wrote:
> On Wed, Jul 31, 2019 at 10:31 PM nokia ceph
> wrote:
> >
> > Thank you Gre
Hi Team,
@vita...@yourcmc.ru , thank you for information and could you please
clarify on the below quires as well,
1. Average object size we use will be 256KB to 512KB , will there be
deferred write queue ?
2. Share the link of existing rocksdb ticket which does 2 write + syncs.
3. Any configurat
On Mon, Aug 5, 2019 at 6:35 PM wrote:
> > Hi Team,
> > @vita...@yourcmc.ru , thank you for information and could you please
> > clarify on the below quires as well,
> >
> > 1. Average object size we use will be 256KB to 512KB , will there be
> > deferred write queue ?
>
> With the default settin
Hi Team,
One of our old customer had Kraken and they are going to upgrade to
Luminous . In the process they also requesting for downgrade procedure.
Kraken used leveldb for ceph-mon data , from luminous it changed to rocksdb
, upgrade works without any issues.
When we downgrade , the ceph-mon doe
e a new old one,
> let it sync, etc.
> Still a bad idea.
>
> Paul
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io
> Tel: +49 89
Hi,
We have a 5 node Luminous cluster on which we see multiple RESETSESSION
messages for OSDs on the last node alone.
's=STATE_CONNECTING_WAIT_CONNECT_REPLY_AUTH pgs=2613 cs=1
l=0).handle_connect_reply connect got RESETSESSION'
We found the below fix for this issue, but not able to identify the
Hi Team,
In ceph 14.2.2 , ceph dashboard does not have set-ssl-certificate .
We are trying to enable ceph dashboard and while using the ssl certificate
and key , it is not working .
cn5.chn5au1c1.cdn ~# ceph dashboard set-ssl-certificate -i dashboard.crt
no valid command found; 10 closest matches:
board.crt
> $ ceph config-key set mgr/dashboard/key -i dashboard.key
>
> The above commands will emit a deprecation warning that you can ignore.
>
> Thanks,
> Ricardo Dias
>
> ____
> From: ceph-users on behalf of nokia
> ceph
> Sent
Hi Team,
With default log settings , the ceph stats will be logged like
cluster [INF] pgmap v30410386: 8192 pgs: 8192 active+clean; 445 TB data,
1339 TB used, 852 TB / 2191 TB avail; 188 kB/s rd, 217 MB/s wr, 1618 op/s
Jewel : on mon logs
Nautilus : on mgr logs
Luminous : not able to view simil
Hi Team,
We have noticed that memory usage of ceph-monitor processes increased by
1GB in 4 days.
We monitored the ceph-monitor memory usage every minute and we can see it
increases and decreases by few 100 MBs at any point; but over time, the
memory usage increases. We also noticed some monitor pr
Hi Team,
There is one 5 node ceph cluster which we have upgraded from Luminous to
Nautilus and everything was going well until yesterday when we noticed that
the ceph osd's are marked down and not recognized by the monitors as
running eventhough the osd processes are running.
We noticed that the
Adding my official mail id
-- Forwarded message -
From: nokia ceph
Date: Fri, Nov 8, 2019 at 3:57 PM
Subject: OSD's not coming up in Nautilus
To: Ceph Users
Hi Team,
There is one 5 node ceph cluster which we have upgraded from Luminous to
Nautilus and everything was
down osds in 'ceph osd tree', and to see
> what happened?
>
> nokia ceph 于2019年11月8日周五 下午6:24写道:
> >
> > Adding my official mail id
> >
> > ------ Forwarded message -
> > From: nokia ceph
> > Date: Fri, Nov 8, 2019 at 3:5
dc44-c17e-42ee-bcd4-d577c0ef4479)
> the osd boot will be ignored if the fsid mismatch
> what do you do before this happen?
>
> nokia ceph 于2019年11月8日周五 下午8:29写道:
> >
> > Hi,
> >
> > Please find the osd.0 which is restarted after the debug_mon is
> increased to
Hi,
Please find the ceph osd tree output in the pastebin
https://pastebin.com/Gn93rE6w
On Fri, Nov 8, 2019 at 7:58 PM huang jun wrote:
> can you post your 'ceph osd tree' in pastebin?
> do you mean the osds report fsid mismatch is from old removed nodes?
>
> nokia ceph 于
The mon log shows that the all mismatch fsid osds are from node
> 10.50.11.45,
> maybe that the fith node?
> BTW i don't found the osd.0 boot message in ceph-mon.log
> do you set debug_mon=20 first and then restart osd.0 process, and make
> sure the osd.0 is restarted.
>
>
&g
s '--debug-mon 1/5'
injectargs:
cn1.chn8be1c1.cdn ~# ceph daemon /var/run/ceph/ceph-mon.cn1.asok config
show|grep debug_mon
"debug_mon": "1/5",
"debug_monc": "0/0",
On Sun, Nov 10, 2019 at 11:05 AM huang jun wrote:
> good, please s
Nov 8 23:39:32 UTC 2018",
"kernel_version": "3.10.0-957.el7.x86_64",
"mem_swap_kb": "0",
"mem_total_kb": "272036636",
"network_numa_unknown_ifaces": "dss-client,dss-private",
"objectstore_nu
Hi Team,
In one of our ceph cluster we observe that there are many slow IOPS in all
our OSD's and most of the latency is happening between two set of
operations which are shown below.
{
"time": "2019-11-12 08:29:58.128669",
"even
Hi,
We have upgraded a 5 node ceph cluster from Luminous to Nautilus and the
cluster was running fine. Yesterday when we tried to add one more osd into
the ceph cluster we find that the OSD is created in the cluster but
suddenly some of the other OSD's started to crash and we are not able to
resta
Hi Team,
We would like to create multiple snapshots inside ceph cluster,
initiate the request from librados client and came across this rados api
rados_ioctx_selfmanaged_snap_set_write_ctx
Can some give us sample code on how to use this api .
Thanks,
Muthu
_
Hi,
When installing Nautilus on a five node cluster, we tried to install one
node first and then the remaining four nodes. After that we saw that the
fifth node is out of quorum and we found that the fsid was different in 5th
node. When we replaced the ceph.conf file from the four nodes to the fif
95 matches
Mail list logo