Hi Dan
Hope to find you ok.
Here goes a suggestion from someone who has been sitting in the side line
for the last 2 years but following stuff as much as possible
Will weight set per pool help?
This is only possible in luminous but according to the docs there is the
possibility to adjust positi
Thanks guys, installing this package did the trick, it works now.
On Mon, Sep 24, 2018 at 8:39 AM Ken Dreyer wrote:
> Hi Alfredo,
>
> I've packaged the latest version in Fedora, but I didn't update EPEL.
> I've submitted the update for EPEL now at
> https://bodhi.fedoraproject.org/updates/FEDO
Good news... :)
After I tried everything. I decide to re-create my MONs from OSD's and
I used the script:
https://paste.ubuntu.com/p/rNMPdMPhT5/
And it worked!!!
I think when 2 server crashed and come back same time some how MON's
confused and the maps just corrupted.
After re-creation all the MO
On 27/09/18 17:18, Dan van der Ster wrote:
Dear Ceph friends,
I have a CRUSH data migration puzzle and wondered if someone could
think of a clever solution.
Consider an osd tree like this:
-2 4428.02979 room 0513-R-0050
-72911.81897 rack RA01
-4917.
I am running luminous and the objects were copied from Isilon with a long
and similar prefix in path like /dir1/dir2/dir3//mm/dd. The objects are
copied to various buckets like bucket_MMDD/dir1/dir2/dir3//mm/dd.
This setup minimize some internal code change when moving from NFS to
objec
I see Filestore symbols on the stack, so the bluestore config doesn’t affect.
And the top frame of the stack hints at a RocksDB issue, and there are a whole
lot of these too:
“2018-09-17 19:23:06.480258 7f1f3d2a7700 2 rocksdb:
[/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILAB
I think I might find something.
When I start an OSD its making High I/O around %95 and the other OSDs
are also triggered and altogether they make same the I/O. This is true
even if when I set noup flag. So all the OSDs are making high I/O when
ever an OSD starts.
I think this is too much. I have
I got pulled away from this for a while. The error in the log is "abort:
Corruption: Snappy not supported or corrupted Snappy compressed block
contents" and the OSD has 2 settings set to snappy by default,
async_compressor_type and bluestore_compression_algorithm. Do either of
these settings affe
I think your objective is to move the data without anyone else
noticing. What I usually do is reduce the priority of the recovery
process as much as possible. Do note this will make the recovery take
a looong time, and will also make recovery from failures slow...
ceph tell osd.* injectargs '--osd_
Dear Ceph friends,
I have a CRUSH data migration puzzle and wondered if someone could
think of a clever solution.
Consider an osd tree like this:
-2 4428.02979 room 0513-R-0050
-72911.81897 rack RA01
-4917.27899 rack RA05
-6917.25500
As of today, there is no such feature in Ceph
Best regards,
On 09/27/2018 04:34 PM, Gaël THEROND wrote:
> Hi folks!
>
> As I’ll soon start to work on a new really large an distributed CEPH
> project for cold data storage, I’m checking out a few features availability
> and status, with the need
Hi all,
I am exporting cephfs using samba. It is much slower over samba than
direct. Anyone know how to speed it up?
Benchmarked using bonnie++ 5 times either directly to cephfs mounted
by kernel (v4.18.6) module:
bonnie++ -> kcephfs
or through a cifs kernel-module-mounted (protocol ve
Hi folks!
As I’ll soon start to work on a new really large an distributed CEPH
project for cold data storage, I’m checking out a few features availability
and status, with the need for deduplication among them.
I found out an interesting video about that from Cephalocon APAC 2018 and a
seven year
If I add on one client a file to the cephfs, that is exported via
ganesha and nfs mounted somewhere else. I can see it in the dir listing
on the other nfs client. But trying to read it gives an Input/output
error. Other files (older ones in the same dir I can read)
Anyone had this also?
nfs
I should not have client I/O right now. All of my VMs are down right
now. There is only a single pool.
Here is my crush map: https://paste.ubuntu.com/p/Z9G5hSdqCR/
Cluster does not recover. After starting OSDs with the specified
flags, OSD up count drops from 168 to 50 with in 24 hours.
Stefan Ko
Quoting by morphin (morphinwith...@gmail.com):
> After 72 hours I believe we may hit a bug. Any help would be greatly
> appreciated.
Is it feasible for you to stop all client IO to the Ceph cluster? At
least until it stabilizes again. "ceph osd pause" would do the trick
(ceph osd unpause would uns
Hello,
I am writing this e-mail about an incident that has started last weekend.
There seems to something wrong with my e-mail. Some of my e-mails did
not reach-out. So I decided to start an new thread here and start from
begining.
One can find the email related e-mail thread
(http://lists.ceph.co
>
> On 27.09.2018, at 15:04, John Spray wrote:
>
> On Thu, Sep 27, 2018 at 11:34 AM Sergey Malinin wrote:
>>
>> Can such behaviour be related to data pool cache tiering?
>
> Yes -- if there's a cache tier in use then deletions in the base pool
> can be delayed and then happen later when the c
On Thu, Sep 27, 2018 at 11:34 AM Sergey Malinin wrote:
>
> Can such behaviour be related to data pool cache tiering?
Yes -- if there's a cache tier in use then deletions in the base pool
can be delayed and then happen later when the cache entries get
expired.
You may find that for a full scan of
Can such behaviour be related to data pool cache tiering?
> On 27.09.2018, at 13:14, Sergey Malinin wrote:
>
> I'm trying alternate metadata pool approach. I double checked that MDS
> servers are down and both original and recovery fs are set not joinable.
>
>
>> On 27.09.2018, at 13:10, Joh
I'm trying alternate metadata pool approach. I double checked that MDS servers
are down and both original and recovery fs are set not joinable.
> On 27.09.2018, at 13:10, John Spray wrote:
>
> On Thu, Sep 27, 2018 at 11:03 AM Sergey Malinin wrote:
>>
>> Hello,
>> Does anybody have experience
On 26/09/2018 12:41, Eugen Block wrote:
Hi,
I'm not sure how the recovery "still works" with the flag "norecover".
Anyway, I think you should unset the flags norecover, nobackfill. Even
if not all OSDs come back up you should allow the cluster to backfill
PGs. Not sure, but unsetting norebalan
On Thu, Sep 27, 2018 at 11:03 AM Sergey Malinin wrote:
>
> Hello,
> Does anybody have experience with using cephfs-data-scan tool?
> Questions I have are how long would it take to scan extents on filesystem
> with 120M relatively small files? While running extents scan I noticed that
> number of
Hello,
Does anybody have experience with using cephfs-data-scan tool?
Questions I have are how long would it take to scan extents on filesystem with
120M relatively small files? While running extents scan I noticed that number
of objects in data pool is decreasing over the time. Is that normal?
T
On Thu, Sep 27, 2018 at 10:16 AM Marc Roos wrote:
>
>
> I have a test cluster and on a osd node I put a vm. The vm is using a
> macvtap on the client network interface of the osd node. Making access
> to local osd's impossible.
>
> the vm of course reports that it cannot access the local osd's. Wh
Hi,
On 09/27/2018 11:15 AM, Marc Roos wrote:
I have a test cluster and on a osd node I put a vm. The vm is using a
macvtap on the client network interface of the osd node. Making access
to local osd's impossible.
the vm of course reports that it cannot access the local osd's. What I
am getting
I have a test cluster and on a osd node I put a vm. The vm is using a
macvtap on the client network interface of the osd node. Making access
to local osd's impossible.
the vm of course reports that it cannot access the local osd's. What I
am getting is:
- I cannot reboot this vm normally, ne
Hi,
I was reading this thread:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-March/008486.html
And I am trying to get better performance in my virtual machines.
These are my RBD settings:
"rbd_cache": "true",
"rbd_cache_block_writes_upfront": "false",
"rbd_cache_max_dirty":
28 matches
Mail list logo