[ceph-users] stat file size is 0

2018-08-13 Thread xiang . dai
I mount cephfs in /cephfs and create a dir in it: [root@test-0 guhailin]# ll -h drwxr-xr-x 1 guhailin ghlclass 0 8月 13 15:01 a scp a file into it: [root@test-0 guhailin]# ls a/ hadoop.tar.gz [root@test-0 guhailin]# pwd /cephfs/user/guhailin [root@test-0 guhailin]# stat a/

[ceph-users] Help needed for debugging slow_requests

2018-08-13 Thread Uwe Sauter
Dear community, TL;DR: Cluster runs good with Kernel 4.13, produces slow_requests with Kernel 4.15. How to debug? I'm running a combined Ceph / KVM cluster consisting of 6 hosts of 2 different kinds (details at the end). The main difference between those hosts is CPU generation (Westmere /

Re: [ceph-users] cephfs - restore files

2018-08-13 Thread John Spray
On Thu, Aug 9, 2018 at 6:34 PM Erik Schwalbe wrote: > > Hi, > > Unfortunately, I deleted a few files and would like to restore them. > For ext4 I would take photorec but that not seem to work for cephfs. > > Is it possible to restore deleted files stored in cephfs? In general, no. There is a sho

Re: [ceph-users] Make a ceph options persist

2018-08-13 Thread John Spray
On Fri, Aug 10, 2018 at 10:40 AM Willem Jan Withagen wrote: > > Hi, > > The manual of dashboard suggests: > ceph config-key set mgr/dashboard/server_addr ${MGR_IP} > > But that command is required after reboot. config-key settings are persistent. The docs are probably just telling you to

Re: [ceph-users] Make a ceph options persist

2018-08-13 Thread Willem Jan Withagen
On 13/08/2018 10:51, John Spray wrote: On Fri, Aug 10, 2018 at 10:40 AM Willem Jan Withagen wrote: Hi, The manual of dashboard suggests: ceph config-key set mgr/dashboard/server_addr ${MGR_IP} But that command is required after reboot. config-key settings are persistent. The docs

Re: [ceph-users] RBD image "lightweight snapshots"

2018-08-13 Thread Konstantin Shalygin
1. Create snapshot of the image we want to backup 2. If there's a previous backup snapshot, export diff and apply it on the backup image 3. If there's no older snapshot, just do a full backup of image So you need incremental backup? Try look to "rbd2qcow2" [1] [1] https://github.com/socketpai

Re: [ceph-users] Optane 900P device class automatically set to SSD not NVME

2018-08-13 Thread vitalif
Hi, Can you benchmark your Optane 900P with `fio -fsync=1 -direct=1 -bs=4k -rw=randwrite -runtime=60`? It's really interesting to see how much iops it will provide for ceph :) -- With best regards, Vitaliy Filippov ___ ceph-users mailing list ceph

Re: [ceph-users] RBD image "lightweight snapshots"

2018-08-13 Thread Bartosz Rabiega
On 08/11/2018 07:56 AM, Paweł Sadowski wrote: On 08/10/2018 06:24 PM, Gregory Farnum wrote: On Fri, Aug 10, 2018 at 4:53 AM, Paweł Sadowsk wrote: On 08/09/2018 04:39 PM, Alex Elder wrote: On 08/09/2018 08:15 AM, Sage Weil wrote: On Thu, 9 Aug 2018, Piotr Dałek wrote: Hello, At OVH we're

Re: [ceph-users] [Jewel 10.2.11] OSD Segmentation fault

2018-08-13 Thread Alexandru Cucu
Hi, Already tried zapping the disk. Unfortunaltely the same segfaults keep me from adding the OSD back to the cluster. I wanted to open an issue on tracker.ceph.com but I can't find the "new issue" button. --- Alex Cucu On Mon, Aug 13, 2018 at 8:24 AM wrote: > > > > Am 3. August 2018 12:03:17

[ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
Hi, Recently, the cluster runs healthy, but I get warning messages everyday: 2018-08-13 17:39:23.682213 [INF] Cluster is now healthy 2018-08-13 17:39:23.682144 [INF] Health check cleared: MDS_CLIENT_RECALL (was: 6 clients failing to respond to cache pressure) 2018-08-13 17:39:23.052022 [INF] MD

Re: [ceph-users] bad crc/signature errors

2018-08-13 Thread Nikola Ciprich
Hi Paul, thanks, I'll give it a try.. do you think this might head to upstream soon? for some reason I can't review comments for this patch on github.. Is some new version of this patch on the way, or can I try to apply this one to latest luminous? thanks a lot! nik On Fri, Aug 10, 2018 at 06

Re: [ceph-users] Optane 900P device class automatically set to SSD not NVME

2018-08-13 Thread Alfredo Deza
On Wed, Aug 1, 2018 at 4:33 AM, Jake Grimmett wrote: > Dear All, > > Not sure if this is a bug, but when I add Intel Optane 900P drives, > their device class is automatically set to SSD rather than NVME. Not sure if we can't really tell apart SSDs from NVMe devices, but you can use the --crush-de

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Wido den Hollander
On 08/13/2018 01:22 PM, Zhenshi Zhou wrote: > Hi, > Recently, the cluster runs healthy, but I get warning messages everyday: > Which version of Ceph? Which version of clients? Can you post: $ ceph versions $ ceph features $ ceph fs status Wido > 2018-08-13 17:39:23.682213 [INF]  Cluster is

[ceph-users] Replicating between two datacenters without decompiling CRUSH map

2018-08-13 Thread Torsten Casselt
Hi, I created a rule with this command: ceph osd crush rule create-replicated rrd default datacenter Since chooseleaf type is 1, I expected it to distribute the copies evenly on two datacenters with six hosts each. For example, six copies would mean a copy on each host. When I test the resultin

Re: [ceph-users] Stale PG data loss

2018-08-13 Thread Surya Bala
Any suggestion on this please Regards Surya Balan On Fri, Aug 10, 2018 at 11:28 AM, Surya Bala wrote: > Hi folks, > > I was trying to test the below case > > Having pool with replication count as 1 and if one osd goes down, then the > PGs mapped to that OSD become stale. > > If the hardware fa

Re: [ceph-users] Optane 900P device class automatically set to SSD not NVME

2018-08-13 Thread Olivier Bonvalet
On a recent Luminous cluster, with nvme*n1 devices, the class is automatically set as "nvme" on "Intel SSD DC P3520 Series" : ~# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 2.15996 root default -9 0.71999 roo

Re: [ceph-users] Stale PG data loss

2018-08-13 Thread Janne Johansson
"Don't run with replication 1 ever". Even if this is a test, it tests something for which a resilient cluster is specifically designed to avoid. As for enumerating what data is missing, it would depend on if the pool(s) had cephfs, rbd images or rgw data in them. When this kind of data loss happe

[ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Dear ceph users, I set up a new cluster: - Debian stretch - ceph 12.2.7 - 3 nodes with mixed mon/osd - 4 hdd  4TB osd per nodes - 2 SSDs per nodes shared among osds for db/wal - each OSD alone in a raid0+WriteBack Inside a VM I get really good writes(200MB/s, 5k iops for direct 4K rand writes),

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Jason Dillaman
Is this a clean (new) cluster and RBD image you are using for your test or has it been burned in? When possible (i.e. it has enough free space), bluestore will essentially turn your random RBD image writes into sequential writes. This optimization doesn't work for random reads unless your read patt

Re: [ceph-users] cephfs kernel client hangs

2018-08-13 Thread Zhenshi Zhou
Hi, Finally, I got a running server with files /sys/kernel/debug/ceph/xxx/ [root@docker27 525c4413-7a08-40ca-9a98-0a6df009025b.client213522]# cat mdsc [root@docker27 525c4413-7a08-40ca-9a98-0a6df009025b.client213522]# cat monc have monmap 2 want 3+ have osdmap 4545 want 4546 have fsmap.user 0 have

Re: [ceph-users] bad crc/signature errors

2018-08-13 Thread Ilya Dryomov
On Mon, Aug 13, 2018 at 2:49 PM Nikola Ciprich wrote: > > Hi Paul, > > thanks, I'll give it a try.. do you think this might head to > upstream soon? for some reason I can't review comments for > this patch on github.. Is some new version of this patch > on the way, or can I try to apply this one

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Le 13/08/2018 à 15:21, Jason Dillaman a écrit : > Is this a clean (new) cluster and RBD image you are using for your > test or has it been burned in? When possible (i.e. it has enough free > space), bluestore will essentially turn your random RBD image writes > into sequential writes. This optimiza

Re: [ceph-users] cephfs kernel client hangs

2018-08-13 Thread Burkhard Linke
Hi, On 08/13/2018 03:22 PM, Zhenshi Zhou wrote: Hi, Finally, I got a running server with files /sys/kernel/debug/ceph/xxx/ [root@docker27 525c4413-7a08-40ca-9a98-0a6df009025b.client213522]# cat mdsc [root@docker27 525c4413-7a08-40ca-9a98-0a6df009025b.client213522]# cat monc have monmap 2 want

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Eugen Block
Hi, Depending on your kernel (memory leaks with CephFS) increasing the mds_cache_memory_limit could be of help. What is your current setting now? ceph:~ # ceph daemon mds. config show | grep mds_cache_memory_limit We had these messages for months, almost every day. It would occur when hour

Re: [ceph-users] cephfs kernel client hangs

2018-08-13 Thread Zhenshi Zhou
Hi Burkhard, I'm sure the user has permission to read and write. Besides, we're not using EC data pools. Now the situation is that any openration to a specific file, the command will hang. Operations to any other files won't hang. Burkhard Linke 于2018年8月13日周一 下午9:42写道: > Hi, > > > On 08/13/2018

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Jason Dillaman
On Mon, Aug 13, 2018 at 9:32 AM Emmanuel Lacour wrote: > Le 13/08/2018 à 15:21, Jason Dillaman a écrit : > > Is this a clean (new) cluster and RBD image you are using for your > > test or has it been burned in? When possible (i.e. it has enough free > > space), bluestore will essentially turn you

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Le 13/08/2018 à 15:55, Jason Dillaman a écrit : > > > > so problem seems located on "rbd" side  ... > > > That's a pretty big apples-to-oranges comparison (4KiB random IO to > 4MiB full-object IO). With your RBD workload, the OSDs will be seeking > after each 4KiB read but w/ your RADOS bench w

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
Hi Eugen, The command shows "mds_cache_memory_limit": "1073741824". And I'll increase the cache size for a try. Thanks Eugen Block 于2018年8月13日周一 下午9:48写道: > Hi, > > Depending on your kernel (memory leaks with CephFS) increasing the > mds_cache_memory_limit could be of help. What is your current

Re: [ceph-users] failing to respond to cache pressure

2018-08-13 Thread Zhenshi Zhou
Hi Wido, The server and client use the same version, 12.2.5. And the command shows: # ceph versions { "mon": { "ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a) luminous (stable)": 3 }, "mgr": { "ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a

Re: [ceph-users] LVM on top of RBD apparent pagecache corruption with snapshots

2018-08-13 Thread Ilya Dryomov
On Mon, Aug 6, 2018 at 8:17 PM Ilya Dryomov wrote: > > On Mon, Aug 6, 2018 at 8:13 PM Ilya Dryomov wrote: > > > > On Thu, Jul 26, 2018 at 1:55 AM Alex Gorbachev > > wrote: > > > > > > On Wed, Jul 25, 2018 at 7:07 PM, Alex Gorbachev > > > wrote: > > > > On Wed, Jul 25, 2018 at 6:07 PM, Alex Go

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Jason Dillaman
On Mon, Aug 13, 2018 at 10:01 AM Emmanuel Lacour wrote: > Le 13/08/2018 à 15:55, Jason Dillaman a écrit : > > > >> >> so problem seems located on "rbd" side ... >> > > That's a pretty big apples-to-oranges comparison (4KiB random IO to 4MiB > full-object IO). With your RBD workload, the OSDs wil

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Le 13/08/2018 à 16:29, Jason Dillaman a écrit : > > For such a small benchmark (2 GiB), I wouldn't be surprised if you are > not just seeing the Filestore-backed OSDs hitting the page cache for > the reads whereas the Bluestore-backed OSDs need to actually hit the > disk. Are the two clusters simil

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Le 13/08/2018 à 16:29, Jason Dillaman a écrit : > > > For such a small benchmark (2 GiB), I wouldn't be surprised if you are > not just seeing the Filestore-backed OSDs hitting the page cache for > the reads whereas the Bluestore-backed OSDs need to actually hit the > disk. Are the two clusters sim

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Emmanuel Lacour
Le 13/08/2018 à 16:44, Emmanuel Lacour a écrit : > Le 13/08/2018 à 16:29, Jason Dillaman a écrit : >> >> >> For such a small benchmark (2 GiB), I wouldn't be surprised if you >> are not just seeing the Filestore-backed OSDs hitting the page cache >> for the reads whereas the Bluestore-backed OSDs n

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-13 Thread Jason Dillaman
On Mon, Aug 13, 2018 at 10:44 AM Emmanuel Lacour wrote: > Le 13/08/2018 à 16:29, Jason Dillaman a écrit : > > > > For such a small benchmark (2 GiB), I wouldn't be surprised if you are not > just seeing the Filestore-backed OSDs hitting the page cache for the reads > whereas the Bluestore-backed

Re: [ceph-users] RBD image "lightweight snapshots"

2018-08-13 Thread Jason Dillaman
On Fri, Aug 10, 2018 at 8:29 AM Sage Weil wrote: > On Fri, 10 Aug 2018, Paweł Sadowski wrote: > > On 08/09/2018 04:39 PM, Alex Elder wrote: > > > On 08/09/2018 08:15 AM, Sage Weil wrote: > > >> On Thu, 9 Aug 2018, Piotr Dałek wrote: > > >>> Hello, > > >>> > > >>> At OVH we're heavily utilizing sn

Re: [ceph-users] bad crc/signature errors

2018-08-13 Thread Nikola Ciprich
Hi Ilya, hmm, OK, I'm not sure now whether this is the bug which I'm experiencing.. I've had read_partial_message / bad crc/signature problem occurance on the second cluster in short period even though we're on the same ceph version (12.2.5) for quite long time (almost since its release), so it'

Re: [ceph-users] RBD journal feature

2018-08-13 Thread Jason Dillaman
On Sun, Aug 12, 2018 at 12:13 AM Glen Baars wrote: > Hello Jason, > > > > Interesting, I used ‘rados ls’ to view the SSDPOOL and can’t see any > objects. Is this the correct way to view the journal objects? > You won't see any journal objects in the SSDPOOL until you issue a write: $ rbd create

Re: [ceph-users] RBD mirroring replicated and erasure coded pools

2018-08-13 Thread Jason Dillaman
On Tue, Jul 31, 2018 at 11:10 AM Ilja Slepnev wrote: > Hi, > > is it possible to establish RBD mirroring between replicated and erasure > coded pools? > I'm trying to setup replication as described on > http://docs.ceph.com/docs/master/rbd/rbd-mirroring/ without success. > Ceph 12.2.5 Luminous >

Re: [ceph-users] LVM on top of RBD apparent pagecache corruption with snapshots

2018-08-13 Thread Alex Gorbachev
On Mon, Aug 13, 2018 at 10:25 AM, Ilya Dryomov wrote: > On Mon, Aug 6, 2018 at 8:17 PM Ilya Dryomov wrote: >> >> On Mon, Aug 6, 2018 at 8:13 PM Ilya Dryomov wrote: >> > >> > On Thu, Jul 26, 2018 at 1:55 AM Alex Gorbachev >> > wrote: >> > > >> > > On Wed, Jul 25, 2018 at 7:07 PM, Alex Gorbachev

[ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-13 Thread Steven Vacaroaia
Hi, I am in the process of deploying CEPPH mimic on a few DELL R620/630 servers These servers have only 8 disk slots and the larger disk they accept is 2 TB Should I share a SSD drive between OS and WAL/DB or ran OS on internal SD cards and dedicated SSD to DB/WAL only ? Another option is to use

Re: [ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-13 Thread thomas
Hi Steven, If you are running OSDs on the SD card, there would be nothing technically stopping this setup, but the main factors against would be the simple endurance and performance of SD cards and the potential fallout when they inevitably fail. If you factor time and maintenance as a cost

Re: [ceph-users] Least impact when adding PG's

2018-08-13 Thread ceph
Am 7. August 2018 18:08:05 MESZ schrieb John Petrini : >Hi All, Hi John, > >Any advice? > I am Not sure but what i would do is to increase the PG Step by Step and always with a value of "Power of two" i.e. 256. Also have a look on the pg_-/pgp_num. One of this should be increased First -

Re: [ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-13 Thread Thomas White
Hi Steven, Just to somewhat clarify my previous post, I mention OSDs in the sense that the OS is installed on the OSD server using the SD card, I would absolutely recommend against using SD cards as the actual OSD media. This of course misses another point, which is for the Mons or other suc

Re: [ceph-users] [Jewel 10.2.11] OSD Segmentation fault

2018-08-13 Thread Brad Hubbard
Jewel is almost EOL. It looks similar to several related issues, one of which is http://tracker.ceph.com/issues/21826 On Mon, Aug 13, 2018 at 9:19 PM, Alexandru Cucu wrote: > Hi, > > Already tried zapping the disk. Unfortunaltely the same segfaults keep > me from adding the OSD back to the clust

[ceph-users] 答复: pg unexpected down on luminous

2018-08-13 Thread 平安科技公共平台开发部文件服务组
Hello guys, We are testing ceph on luminous. We found pg unexpected down, when osd down. In one pg map 14.1f0,we got 3 osds 4,6,0. When we do up/down osds in this order, pg will down at last. All step are waiting for peering over. We found this happened only on luminous ,we are tested on jewel an

[ceph-users] can we get create time ofsnap

2018-08-13 Thread hnuzhoulin2
hi,guyswe have much snaps,and we want to clear.but can not know the create time of these snaps.I know luminous have this feature default.but for jewel and hammer,have some hack way to get the create time?or it is just impossible.Thanks __

Re: [ceph-users] Least impact when adding PG's

2018-08-13 Thread Janne Johansson
Den mån 13 aug. 2018 kl 23:30 skrev : > > > Am 7. August 2018 18:08:05 MESZ schrieb John Petrini < > jpetr...@coredial.com>: > >Hi All, > > Hi John, > > > > >Any advice? > > > > I am Not sure but what i would do is to increase the PG Step by Step and > always with a value of "Power of two" i.e. 25