Hi,
We are running the following radosgw( luminous 12.2.8) replications
scenario.
1) We have 2 clusters, each running a radosgw, Cluster1 defined as master,
and Cluster2 as slave.
2) We create a number of bucket with objects via master and slave
3) We shutdown the Cluster1
4) We execute failover on
I’m using krbd to map a rbd device to a VM, it appears when the device is
mounted, reboot OS will hung for more than 7min, in baremetal case, it could be
more than 15min, even using the latest kernel 5.0.0, the problem still occurs.
Here are the console logs with 4.15.18 kernel and mimic rbd clie
Hello.
I have a question about `ceph pg {pg.num} mark_unfound_lost delete`.
Will this only delete objects which are unfound, or the whole PG which
you put in as an argument?
Objects (oid's) which i can see with `ceph pg {pg.num} list_missing`?
So in the case bellow, would it remove the object
"rbd_
dear all,
i have a luminious cluster with tunables profile "hammer" - now all my
hammer clients are gone and i could raise the tunables level to "jewel".
is there any good way to predict the data movement caused by such a
config change?
br
wolfgang
smime.p7s
Description: S/MIME Cryptograp
Den tis 22 jan. 2019 kl 00:50 skrev Brian Topping :
> > I've scrounged up 5 old Atom Supermicro nodes and would like to run them
> > 365/7 for limited production as RBD with Bluestore (ideally latest 13.2.4
> > Mimic), triple copy redundancy. Underlying OS is a Debian 9 64 bit, minimal
> > insta
On Tue, Jan 22, 2019 at 10:49 AM Albert Yue wrote:
>
> Hi Yan Zheng,
>
> In your opinion, can we resolve this issue by move MDS to a 512GB or 1TB
> memory machine?
>
The problem is from client side, especially clients with large memory.
I don't think enlarge mds cache size is good idea. you can
Hello - If an OSD shown as down and but its still "in" state..what
will happen with write/read operations on this down OSD?
Thanks
Swami
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Hi Thomas,
What is the best practice for creating pools & images?
Should I create multiple pools, means one pool per database?
Or should I create a single pool "backup" and use namespace when writing
data in the pool?
I don't think one pool per DB is reasonable. If the number of DBs
increase
Hi,
we want upgrade our ceph clusters from jewel to luminous. And also want
to migrate the osds to ceph-volume described in
http://docs.ceph.com/docs/luminous/ceph-volume/simple/scan/#ceph-volume-simple-scan
The clusters are running since dumpling and are setup with dmcrypt.
Since dumpling there
Hi,
On 22/01/2019 10:02, M Ranga Swami Reddy wrote:
> Hello - If an OSD shown as down and but its still "in" state..what
> will happen with write/read operations on this down OSD?
It depends ;-)
In a typical 3-way replicated setup with min_size 2, writes to placement
groups on that OSD will stil
hi,
at some time, as cache pressure or caps release failure, client apps mount
got stuck.
my use case is in kubernetes cluster and automatic kernel client mount in
nodes.
is anyone faced with same issue or has related solution?
Brs___
ceph-use
On Tue, Jan 22, 2019 at 6:45 AM Manuel Lausch wrote:
>
> Hi,
>
> we want upgrade our ceph clusters from jewel to luminous. And also want
> to migrate the osds to ceph-volume described in
> http://docs.ceph.com/docs/luminous/ceph-volume/simple/scan/#ceph-volume-simple-scan
>
> The clusters are runn
Hi Zheng,
We also just saw this today and got a bit worried.
Should we change to:
diff --git a/src/mds/CInode.cc b/src/mds/CInode.cc
index e8c1bc8bc1..e2539390fb 100644
--- a/src/mds/CInode.cc
+++ b/src/mds/CInode.cc
@@ -2040,7 +2040,7 @@ void CInode::finish_scatter_gather_update(int type)
Hi,
my use case for Ceph is providing a central backup storage.
This means I will backup multiple databases in Ceph storage cluster.
This is my question:
What is the best practice for creating pools & images?
Should I create multiple pools, means one pool per database?
Or should I create a single
Your "mon" cap should be "profile rbd" instead of "allow r" [1].
[1]
http://docs.ceph.com/docs/master/rbd/rados-rbd-cmds/#create-a-block-device-user
On Mon, Jan 21, 2019 at 9:05 PM ST Wong (ITSC) wrote:
>
> Hi,
>
> > Is this an upgraded or a fresh cluster?
> It's a fresh cluster.
>
> > Does cli
Hi,
I've got a couple of PGs which are stuck in backfill_toofull, but none
of them are actually full.
"up": [
999,
1900,
145
],
"acting": [
701,
1146,
1880
],
"backfill_targets": [
"145",
"999",
"1900"
],
"acting_recovery_backfill": [
"145",
Hi,
Ceph's pool are meant to let you define specific engineering rules
and/or application (rbd, cephfs, rgw)
They are not designed to be created in a massive fashion (see pgs etc)
So, create a pool for each engineering ruleset, and store your data in them
For what is left of your project, I believ
Hi Marc,
My point was that there was no way to do that for a kernel mount except
from the client that consumes the mounted RBDs.
Mohamad
On 1/21/19 4:29 AM, Marc Roos wrote:
>
> Hi Mohamad, How do you do that client side, I am having currently two
> kernel mounts?
>
>
>
>
>
> -Original Mes
On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster wrote:
>
> Hi Zheng,
>
> We also just saw this today and got a bit worried.
> Should we change to:
>
What is the error message (on stray dir or other dir)? does the
cluster ever enable multi-acitive mds?
> diff --git a/src/mds/CInode.cc b/src/mds/
On Tue, Jan 22, 2019 at 3:33 PM Yan, Zheng wrote:
>
> On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster wrote:
> >
> > Hi Zheng,
> >
> > We also just saw this today and got a bit worried.
> > Should we change to:
> >
>
> What is the error message (on stray dir or other dir)? does the
> cluster eve
Hi.
We're currently co-locating our mons with the head node of our Hadoop
installation. That may be giving us some problems, we dont know yet, but
thus I'm speculation about moving them to dedicated hardware.
It is hard to get specifications "small" engough .. the specs for the
mon is where we us
My backup client is using librados.
I understand that defining a pool for the same application is recommended.
However this would not answer my other questions:
How can I identify a backup created by client A that I want to restore
on another client Z?
I mean typically client A would write a backu
AFAIK, the only AAA available with librados works on a pool granularity
So, if you create a ceph user with access to your pool, he will get
access to all the content stored in this pool
If you want to use librados for your use case, you will need to
implement, on your code, the application logic r
Thanks all for the great advices and inputs.
Regarding Mykola's suggestion to use Read-Only snapshots, what is the
overhead of creating these snapshots? I assume these are copy-on-write
snapshots, so there's no extra space consumed except for the metadata?
Thanks,
Shridhar
On Fri, 18 Jan 2019 a
I got one again
[] wait_on_page_bit_killable+0x83/0xa0
[] __lock_page_or_retry+0xb2/0xc0
[] filemap_fault+0x3b7/0x410
[] ceph_filemap_fault+0x13c/0x310 [ceph]
[] __do_fault+0x4c/0xc0
[] do_read_fault.isra.42+0x43/0x130
[] handle_mm_fault+0x6b1/0x1040
[] __do_page_fault+0x154/0x450
[] do_page_fau
If you use librados directly it's up to you to ensure you can identify your
objects. Generally RADOS stores objects and not files so when you provide
your object ids you need to come up with a convention so you can correctly
identify them. If you need to provide meta data (i.e. a list of all
existi
On Wed, Jan 23, 2019 at 5:50 AM Marc Roos wrote:
>
>
> I got one again
>
> [] wait_on_page_bit_killable+0x83/0xa0
> [] __lock_page_or_retry+0xb2/0xc0
> [] filemap_fault+0x3b7/0x410
> [] ceph_filemap_fault+0x13c/0x310 [ceph]
> [] __do_fault+0x4c/0xc0
> [] do_read_fault.isra.42+0x43/0x130
> [] handl
On Tue, Jan 22, 2019 at 8:24 PM renjianxinlover wrote:
>
> hi,
>at some time, as cache pressure or caps release failure, client apps mount
> got stuck.
>my use case is in kubernetes cluster and automatic kernel client mount in
> nodes.
>is anyone faced with same issue or has related
On Tue, Jan 22, 2019 at 10:42 PM Dan van der Ster wrote:
>
> On Tue, Jan 22, 2019 at 3:33 PM Yan, Zheng wrote:
> >
> > On Tue, Jan 22, 2019 at 9:08 PM Dan van der Ster
> > wrote:
> > >
> > > Hi Zheng,
> > >
> > > We also just saw this today and got a bit worried.
> > > Should we change to:
> >
But with enough memory on MDS, I can just cache all metadata into memory.
Right now there are around 500GB metadata in the ssd. So this is not enough?
On Tue, Jan 22, 2019 at 5:48 PM Yan, Zheng wrote:
> On Tue, Jan 22, 2019 at 10:49 AM Albert Yue
> wrote:
> >
> > Hi Yan Zheng,
> >
> > In your o
On Wed, Jan 23, 2019 at 10:02 AM Albert Yue wrote:
>
> But with enough memory on MDS, I can just cache all metadata into memory.
> Right now there are around 500GB metadata in the ssd. So this is not enough?
>
mds needs to tracking lots of extra information for each object. For
500G metadata, md
Thanks for reply.
If the OSD represents the primary one for a PG, then all IO will be
stopped..which may lead to application failure..
On Tue, Jan 22, 2019 at 5:32 PM Matthew Vernon wrote:
>
> Hi,
>
> On 22/01/2019 10:02, M Ranga Swami Reddy wrote:
> > Hello - If an OSD shown as down and but it
32 matches
Mail list logo