On Fri, Jun 22, 2018 at 10:44 PM Gregory Farnum wrote:
>
> On Fri, Jun 22, 2018 at 6:22 AM Sergey Malinin wrote:
>>
>> From
>> http://docs.ceph.com/docs/mimic/rados/troubleshooting/troubleshooting-pg/ :
>>
>> "Now 1 knows that these object exist, but there is no live ceph-osd who has
>> a copy.
Hi all,
We have a Luminous 12.2.5 cluster, running entirely just CephFS with 1 active
and 1 standby MDS. The active MDS crashed and now won't start again with this
same error:
###
0> 2018-06-25 16:11:21.136203 7f01c2749700 -1
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_6
Hi Jason,
your guesses were correct. Thank you for your support.
Just in case, someone else stumbles upon this thread, some more links:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2017-September/020722.html
http://docs.ceph.com/docs/luminous/rados/operations/user-management/#authorization
So my colleague Sean Crosby and I were looking through the logs (with debug mds
= 10) and found some references just before the crash to inode number. We
converted it from HEX to decimal and got something like 109953*5*627776 (last
few digits not necessarily correct). We set one digit up i.e to
Hi Brad,
here is the output:
--
root@arh-ibstorage1-ib:/home/andrei# ceph --debug_ms 5 --debug_auth 20 pg 18.2
query
2018-06-25 10:59:12.100302 7fe23eaa1700 2 Event(0x7fe2400e0140 nevent=5000
time_id=1).set_owner idx=0 owner=140609690670848
2018-06-25 10:59:12.100398 7fe23e2a0700
Hey Ceph people,
need advise on how to move a ceph-cluster from one datacenter to another
without any downtime :)
DC 1:
3 dedicated MON-Server (also MGR on this Servers)
4 dedicated OSD-Server (3x12 OSD, 1x 23 OSDs)
3 Proxmox Nodes with connection to our Ceph-Storage (not managed from
proxmo
Hi All,
I've been using the balancer module in crush-compat mode for quite a while
now and want to switch to upmap mode since all my clients are now luminous
(v12.2.5)
i've reweighted the compat weight-set back to as close as the original
crush weights using 'ceph osd crush reweight-compat'
Befo
Hi,
On 06/22/2018 08:06 AM, dave.c...@dell.com wrote:
I saw these statement from this link (
http://docs.ceph.com/docs/master/rados/operations/crush-map/ ), it that the
reason which leads to the warning?
" This, combined with the default CRUSH failure domain, ensures that replicas or
erasur
Hi,
On 06/20/2018 07:20 PM, David Turner wrote:
We originally used pacemaker to move a VIP between our RGWs, but ultimately
decided to go with an LB in front of them. With an LB you can utilize both
RGWs while they're up, but the LB will shy away from either if they're down
until the check sta
Hello everybody,
I am thinking about making a production three node Ceph cluster with 3x
1.2TB Intel SSD DC P3520 PCIe storage devices. 10.8 (7.2TB 66% for
production)
I am not planning on a journal on a separate ssd. I assume there is no
advantage of this when using pcie storage?
Network
Good Morning,
After removing roughly 20-some rbd shapshots, one of my OSD's has begun
flapping.
ERROR 1
2018-06-25 06:46:39.132257 a0ce2700 -1 osd.8 pg_epoch: 44738 pg[4.e8( v
44721'485588 (44697'484015,44721'485588] local-lis/les=44593/44595
n=2972 ec=9422/9422 lis/c 44593/44593
+Paul
On Mon, Jun 25, 2018 at 5:14 AM, Christian Zunker
wrote:
> Hi Jason,
>
> your guesses were correct. Thank you for your support.
>
> Just in case, someone else stumbles upon this thread, some more links:
> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2017-September/020722.html
> http
Hi all,
We're planning the migration of a VMWare 5.5 cluster backed by a EMC
VNXe 3200 storage appliance to Proxmox.
The VNXe has about 3 year of warranty left and half the disks
unprovisioned, so the current plan is to use the same VNXe for Proxmox
storage. After warranty expires we'll most
Hi,
Is there any information you'd like to grab off this OSD? Anything I
can provide to help you troubleshoot this?
I ask, because if not, I'm going to reformat / rebuild this OSD (unless
there is a faster way to repair this issue).
Thanks,
Dyweni
On 2018-06-25 07:30, Dyweni - Ceph-User
Hello,
When I mount rbd image with -o queue_depth=1024 I can see much improvement,
generally on writes(random write improvement from 3k IOPS on standard
queue_depth to 24k IOPS on queue_depth=1024).
But is there any way to attach rbd disk to KVM instance with custom
queue_depth? I can't find any
I should be able to answer this question for you if you can supply the
output of the following commands. It will print out all of your pool names
along with how many PGs are in that pool. My guess is that you don't have
a power of 2 number of PGs in your pool. Alternatively you might have
multip
Interesing...
Can I see the output of "ceph auth list" and can you test whether you
can query any other pg that has osd.21 as its primary?
On Mon, Jun 25, 2018 at 8:04 PM, Andrei Mikhailovsky wrote:
> Hi Brad,
>
> here is the output:
>
> --
>
> root@arh-ibstorage1-ib:/home/andrei# ce
Hi David,
I am afraid I can't run the command you provide now,because I tried to
remove another osd on that host to see if it would make the data distribution
even and it did.
The pg number of my pools are at power of 2.
Below is from my note before removed another osd:
pool 3
You have 2 different pools. PGs in each pool are going to be a different
size. It's like saying 12x + 13y should equal 2x + 23y because they each
have 25 X's and Y's. Having equal PG counts on each osd is only balanced if
you have a single pool or have a case where all PGs are identical in size.
T
Hi David,
I am sure most(if not all) data are in one pool.
rbd_pool is only for omap for EC rbd.
ceph df:
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
427T 100555G 329T 77.03
POOLS:
NAMEID USED %USED MAX AVAIL OBJECTS
ec_rbd_po
If you look at ceph pg dump, you'll see the size ceph believes each PG is.
>From your ceph df, your PGs for the rbd_pool will be almost zero. So if you
have an osd with 6 of those PGs and another with none of them, but both
osds have the same number of PGs overall... The osd with none of them will
Hi,
Is it normal that I deleted files from the cephfs and ceph didn't
delete the back objects a day later? Until I restart the mds deamon
then it started to release the storage space.
I noticed the doc(http://docs.ceph.com/docs/mimic/dev/delayed-delete/)
says the file is marked as deleted on the
This is the formated pg dump result:
https://pasteboard.co/HrBZv3s.png
You can see the pg distribution of each pool on each osd is fine.
2018-06-26
shadow_lin
发件人:David Turner
发送时间:2018-06-26 10:32
主题:Re: Re: Re: [ceph-users] Uneven data distribution with even pg distribution
after rebala
Hi:
We are using ceph on infiniband and configure it with default
configuration. The ms_type is async + posix. I see there are 3 kinds of
types. Which one is the most stable and the best performance ? Which one do
you suggest shuold I use in production ?
__
24 matches
Mail list logo