Re: [ceph-users] Can't mount Cephfs

2015-08-27 Thread Andrzej Łukawski
Hi, I ran cephfs-journal-tool to inspect journal 12 hours ago - it's still running. Or... it didn't crush yet, although I don't see any output from it. Is it normal behaviour? Thanks for help. Andrzej W dniu 2015-08-26 o 15:49, Gregory Farnum pisze: There is a cephfs-journal-tool that I bel

Re: [ceph-users] ceph monitoring with graphite

2015-08-27 Thread Dan van der Ster
On Wed, Aug 26, 2015 at 6:52 PM, John Spray wrote: > On Wed, Aug 26, 2015 at 3:33 PM, Dan van der Ster wrote: >> Hi Wido, >> >> On Wed, Aug 26, 2015 at 10:36 AM, Wido den Hollander wrote: >>> I'm sending pool statistics to Graphite >> >> We're doing the same -- stripping invalid chars as needed

Re: [ceph-users] Why are RGW pools all prefixed with a period (.)?

2015-08-27 Thread Wido den Hollander
On 08/26/2015 05:17 PM, Yehuda Sadeh-Weinraub wrote: > On Wed, Aug 26, 2015 at 6:26 AM, Gregory Farnum wrote: >> On Wed, Aug 26, 2015 at 9:36 AM, Wido den Hollander wrote: >>> Hi, >>> >>> It's something which has been 'bugging' me for some time now. Why are >>> RGW pools prefixed with a period? >

[ceph-users] question from a new cepher about bucket

2015-08-27 Thread Duanweijun
Hi friends: I put a file(ceph_0.94.2-1.tar.gz, size 23812K) to ceph: ceph@node110:~$ s3cmd ls 2015-08-27 06:13 s3://bkt-key ceph@node110:~$ s3cmd put ceph_0.94.2-1.tar.gz s3://bkt-key WARNING: Module python-magic is not available. Guessing MIME types based on file extensions. ceph_0.94.2-1.tar.

Re: [ceph-users] ceph monitoring with graphite

2015-08-27 Thread Wido den Hollander
On 08/26/2015 04:33 PM, Dan van der Ster wrote: > Hi Wido, > > On Wed, Aug 26, 2015 at 10:36 AM, Wido den Hollander wrote: >> I'm sending pool statistics to Graphite > > We're doing the same -- stripping invalid chars as needed -- and I > would guess that lots of people have written similar json

Re: [ceph-users] ceph monitoring with graphite

2015-08-27 Thread John Spray
On Thu, Aug 27, 2015 at 10:37 AM, Wido den Hollander wrote: > On 08/26/2015 04:33 PM, Dan van der Ster wrote: >> Hi Wido, >> >> On Wed, Aug 26, 2015 at 10:36 AM, Wido den Hollander wrote: >>> I'm sending pool statistics to Graphite >> >> We're doing the same -- stripping invalid chars as needed -

Re: [ceph-users] Can't mount Cephfs

2015-08-27 Thread John Spray
On Thu, Aug 27, 2015 at 9:33 AM, Andrzej Łukawski wrote: > Hi, > > I ran cephfs-journal-tool to inspect journal 12 hours ago - it's still > running. Or... it didn't crush yet, although I don't see any output from it. > Is it normal behaviour? Your cluster is severely damaged, at the RADOS level.

Re: [ceph-users] Why are RGW pools all prefixed with a period (.)?

2015-08-27 Thread Abhishek L
On Thu, Aug 27, 2015 at 3:01 PM, Wido den Hollander wrote: > On 08/26/2015 05:17 PM, Yehuda Sadeh-Weinraub wrote: >> On Wed, Aug 26, 2015 at 6:26 AM, Gregory Farnum wrote: >>> On Wed, Aug 26, 2015 at 9:36 AM, Wido den Hollander wrote: Hi, It's something which has been 'bugging' me

Re: [ceph-users] Question regarding degraded PGs

2015-08-27 Thread Gregory Farnum
On Thu, Aug 27, 2015 at 2:54 AM, Goncalo Borges wrote: > Hey guys... > > 1./ I have a simple question regarding the appearance of degraded PGs. > First, for reference: > > a. I am working with 0.94.2 > > b. I have 32 OSDs distributed in 4 servers, meaning that I have 8 OSD per > server. > > c. Our

Re: [ceph-users] Can't mount Cephfs

2015-08-27 Thread Gregory Farnum
On Thu, Aug 27, 2015 at 11:11 AM, John Spray wrote: > On Thu, Aug 27, 2015 at 9:33 AM, Andrzej Łukawski > wrote: >> Hi, >> >> I ran cephfs-journal-tool to inspect journal 12 hours ago - it's still >> running. Or... it didn't crush yet, although I don't see any output from it. >> Is it normal beh

Re: [ceph-users] Testing CephFS

2015-08-27 Thread Gregory Farnum
On Mon, Aug 24, 2015 at 12:53 PM, Simon Hallam wrote: > The clients are: > [root@gridnode50 ~]# uname -a > Linux gridnode50 4.0.8-200.fc21.x86_64 #1 SMP Fri Jul 10 21:09:54 UTC 2015 > x86_64 x86_64 x86_64 GNU/Linux > [root@gridnode50 ~]# ceph -v > ceph version 0.80.10 (ea6c958c38df1216bf95c927f1

[ceph-users] How to back up RGW buckets or RBD snapshots

2015-08-27 Thread Somnath Roy
Hi, I wanted to know how RGW users are backing up the bucket contents , so that in the disaster scenario user can recreate the setup. I know there is geo replication support but it could be an expensive proposition. I wanted to know if there is any simple solution like plugging in traditional ba

[ceph-users] v0.94.3 Hammer released

2015-08-27 Thread Sage Weil
This Hammer point release fixes a critical (though rare) data corruption bug that could be triggered when logs are rotated via SIGHUP. It also fixes a range of other important bugs in the OSD, monitor, RGW, RGW, and CephFS. All v0.94.x Hammer users are strongly encouraged to upgrade. UPGRADING

[ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Christoph Adomeit
Hello Ceph Users, yesterday I had a defective Gbic in 1 node of my 10 node ceph cluster. The Gbic was working somehow but had 50% packet-loss. Some packets went through, some did not. What happend that the whole cluster did not service requests in time, there were lots of timeouts and so on un

[ceph-users] Disk/Pool Layout

2015-08-27 Thread German Anders
Hi all, I'm planning to deploy a new Ceph cluster with IB FDR 56Gb/s and I've the following HW: *3x MON Servers:* 2x Intel Xeon E5-2600@v3 8C 256GB RAM 1xIB FRD ADPT-DP (two ports for PUB network) 1xGB ADPT-DP Disk Layout: SOFT-RAID: SCSI1 (0,0,0) (sda) - 120.0 GB ATA IN

[ceph-users] 1 hour until Ceph Tech Talk

2015-08-27 Thread Patrick McGarry
Just a reminder that our Performance Ceph Tech Talk with Mark Nelson will be starting in 1 hour. If you are unable to attend there will be a recording posted on the Ceph YouTube channel and linked from the page at: http://ceph.com/ceph-tech-talks/ -- Best Regards, Patrick McGarry Director Ce

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Jan Schermer
Some comments inline. A lot of it depends on your workload, but I'd say you almost certainly need higher-grade SSDs. You can save money on memory. What will be the role of this cluster? VM disks? Object storage? Streaming?... Jan > On 27 Aug 2015, at 17:56, German Anders wrote: > > Hi all, >

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread German Anders
Hi Jan, Thanks for responding the email, regarding the cluster usage, we are going to used it for non-relational databases, Cassandra, mongoDBs and other apps, so we need that this cluster response well to intense io apps, it's going to be connected to HP enclosures with IB FDR also, and mapped

Re: [ceph-users] Ceph Day Raleigh Cancelled

2015-08-27 Thread Patrick McGarry
As far as I know the only feeds we have available are for the blog and the planet. On Wed, Aug 26, 2015 at 11:59 PM, Matt Taylor wrote: > On that note in regards to watching /cephdays for details, the RSS feed > 404's! > > http://ceph.com/cephdays/feed/ > > Regards, > Matt. > > On 27/08/2015 02:

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Jan Schermer
In that case you need fair IOPS and high throughput. Go with S3610 or the Samsungs (or something else that people there can recommend, but for the love of god don't save on drives :)). It's easier to stick to one type of drives and not complicate things. I would also recommend you add one stora

[ceph-users] Hammer for Production?

2015-08-27 Thread Quentin Hartman
I'm currently running Giant in my cluster, and there are a number of things in Hammer that look promising. Today's release announcement for Hammer reminded me to go take a look at it, and I'm getting mixed information about whether or not it is considered safe production yet or not. The main ceph.c

Re: [ceph-users] Hammer for Production?

2015-08-27 Thread Ian Colle
Quentin, Red Hat Ceph Storage 1.3 is based upon Hammer. I guess you can take away from that that we at Red Hat think it's production ready :-) Ian On Thu, Aug 27, 2015 at 10:30 AM, Quentin Hartman < qhart...@direwolfdigital.com> wrote: > I'm currently running Giant in my cluster, and there are

[ceph-users] Sage doing Reddit AMA 02 Sep @ 2p EDT

2015-08-27 Thread Patrick McGarry
Hey cephers, Just wanted to start sharing this so that people could help us generate interest. Sage will be answering questions via a Reddit AMA on Wednesday 02 September from 2-4p EDT. You can see him in the /r/iama calender in the margin on the main page: https://www.reddit.com/r/iama If anyon

Re: [ceph-users] Sage doing Reddit AMA 02 Sep @ 2p EDT

2015-08-27 Thread Ian Colle
Add to your Gcal here: https://www.google.com/calendar/render?eid=Yzh0cDdiOWVsYjVyZXBlZmVocjAxdTNrMzggYW1hdmVyaWZ5QG0&ctz=America/New_York&pli=1&sf=true&output=xml#eventpage_6 Ian > On Aug 27, 2015, at 10:36, Patrick McGarry wrote: > > Hey cephers, > > Just wanted to start sharing this so that

Re: [ceph-users] Hammer for Production?

2015-08-27 Thread Quentin Hartman
Well, assuming you guys are cleaving to the normal Redhat ethos, then that would be a pretty ringing endorsement for production readiness. Thanks! QH On Thu, Aug 27, 2015 at 11:35 AM, Ian Colle wrote: > Quentin, > > Red Hat Ceph Storage 1.3 is based upon Hammer. I guess you can take away > fro

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread German Anders
I already had the SC3510... so I would stick with those :S...the SC 3700 were a little bit high on price and the manager didn't approved that... not my decision...unfortunately. Also I see very hard to add at the moment another node,.. so maybe I can start with 6 OSD instead of 8, and leave those

Re: [ceph-users] shutdown primary monitor, status of the osds on it will not change, and command like 'rbd create xxx' would block

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 The OSDs should be marked down within about 30 seconds. Can you provide additional information such as ceph version and the ceph.conf file. Thanks, - Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1

Re: [ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 This has been discussed a few times. The consensus seems to be to make sure error rates of NICs or other such metrics are included in your monitoring solution. It would also be good to preform periodic network tests like a full size ping with nofrag

Re: [ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Jan Schermer
Don't kick out the node, just deal with it gracefully and without interruption... if the IO reached the quorum number of OSDs then there's no need to block anymore, just queue it. Reads can be mirrored or retried (much quicker, because making writes idempotent, ordered and async is pretty hard a

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 On Thu, Aug 27, 2015 at 10:25 AM, Jan Schermer wrote: > Some comments inline. > A lot of it depends on your workload, but I'd say you almost certainly need > higher-grade SSDs. You can save money on memory. > > What will be the role of this clust

Re: [ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 I know writing to min_size as sync and size-min_size as async has been discussed before and would help here. From what I understand required a lot of code changes and goes against the strong consistency model of Ceph. I'm not sure if it will be imple

[ceph-users] Doubt regarding cephfs in documentation

2015-08-27 Thread Carlos Raúl Laguna
Hello everyone, Reading the documentation about cephfs i found this notice in the "Create a Ceph filesystem" section "The Ceph command line now includes commands for creating and removing file systems, but at present only one filesystem may exist at a time." I am fairly new to ceph, and was tryin

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Jan Schermer
> On 27 Aug 2015, at 20:57, Robert LeBlanc wrote: > > -BEGIN PGP SIGNED MESSAGE- > Hash: SHA256 > > > > > On Thu, Aug 27, 2015 at 10:25 AM, Jan Schermer wrote: >> Some comments inline. >> A lot of it depends on your workload, but I'd say you almost certainly need >> higher-grade SSD

Re: [ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Jan Schermer
Well, there's no other way to get reliable performance and SLAs compared to traditional storage when what you work with is commodity hardware in a mesh-y configuration. And we do like the idea of killing the traditional storage, right? I think 80s called already and wanted their SAN back... Jan

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 On Thu, Aug 27, 2015 at 1:13 PM, Jan Schermer wrote: > >> On 27 Aug 2015, at 20:57, Robert LeBlanc wrote: >> >> -BEGIN PGP SIGNED MESSAGE- >> Hash: SHA256 >> >> >> >> >> On Thu, Aug 27, 2015 at 10:25 AM, Jan Schermer wrote: >>> Some comme

Re: [ceph-users] Defective Gbic brings whole Cluster down

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 +1 :) - Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 On Thu, Aug 27, 2015 at 1:16 PM, Jan Schermer wrote: Well, there's no other way to get reliable performance and SLAs compared to traditiona

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread German Anders
Thanks a lot Robert and Jan for the comments about the available and possible disk layouts. Is there any advice from the point of view of configuration? any tunable parameters, crush algorithm? Thanks a lot, Best regards, *German* 2015-08-27 16:37 GMT-03:00 Robert LeBlanc : > -BEGIN PGP SI

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 I suggest setting logging to 0/5 on everything. Depending on your desire for reliability and availability, you may want to change your pool min_size/size to 2/4 and adjust your CRUSH map to include rack. Then instruct CRUSH to place two copies in eac

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Jan Schermer
> On 27 Aug 2015, at 21:37, Robert LeBlanc wrote: > > -BEGIN PGP SIGNED MESSAGE- > Hash: SHA256 > > > On Thu, Aug 27, 2015 at 1:13 PM, Jan Schermer wrote: > > > >> On 27 Aug 2015, at 20:57, Robert LeBlanc wrote: > >> > >> -BEGIN PGP SIGNED MESSAGE- > >> Hash: SHA256 > >> > >>

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Jan Schermer
Just two little notes here :) below > On 27 Aug 2015, at 21:58, Robert LeBlanc wrote: > > -BEGIN PGP SIGNED MESSAGE- > Hash: SHA256 > > I suggest setting logging to 0/5 on everything. Depending on your desire for > reliability and availability, you may want to change your pool min_size

Re: [ceph-users] Rados: Undefined symbol error

2015-08-27 Thread Aakanksha Pudipeddi-SSI
Hello Brad, Thank you for your response. Looks like the command is undefined. U _ZN5Mutex4LockEb U _ZN5Mutex6UnlockEv U _ZN5MutexC1ERKSsbbbP11CephContext U _ZN5MutexD1Ev Thanks, Aakanksha -Original Message- From: Brad Hubbard

Re: [ceph-users] Doubt regarding cephfs in documentation

2015-08-27 Thread John Spray
On Thu, Aug 27, 2015 at 8:09 PM, Carlos Raúl Laguna wrote: > Hello everyone, > > Reading the documentation about cephfs i found this notice in the "Create a > Ceph filesystem" section > "The Ceph command line now includes commands for creating and removing file > systems, but at present only one f

[ceph-users] RAM usage only very slowly decreases after cluster recovery

2015-08-27 Thread Chad William Seys
Hi all, It appears that OSD daemons only very slowly free RAM after an extended period of an unhealthy cluster (shuffling PGs around). Prior to a power outage (and recovery) around July 25th, the amount of RAM used was fairly constant, at most 10GB (out of 24GB). You can see in the attached P

Re: [ceph-users] Doubt regarding cephfs in documentation

2015-08-27 Thread Carlos Raúl Laguna
Wow i do this every other day with nfs, i can't evenThanks for your answer. Regards 2015-08-27 16:52 GMT-04:00 John Spray : > On Thu, Aug 27, 2015 at 8:09 PM, Carlos Raúl Laguna > wrote: > > Hello everyone, > > > > Reading the documentation about cephfs i found this notice in the > "Create a

Re: [ceph-users] Disk/Pool Layout

2015-08-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 On Thu, Aug 27, 2015 at 1:59 PM, Jan Schermer wrote: > S3500 is faster than S3700? I can compare 3700 x 3510 x 3610 tomorrow but I'd > be very surprised if the S3500 had a _sustained_ throughput better than 36xx > or 37xx. Were you comparing that

[ceph-users] Introducing NodeFabric - for turnkey Ceph deployments

2015-08-27 Thread Andres Toomsalu
Hi All, I would like to introduce NodeFabric - a custom CentOS Atomic Host image - which enables turnkey Ceph deployments. It can be run on any virtual or physical infrastructure: Amazon EC2 cloud, Openstack and VMware private clouds, on top of popular hypervisors (VirtualBox, Parallels, KVM,

Re: [ceph-users] Question regarding degraded PGs

2015-08-27 Thread Goncalo Borges
Hey Greg... Thanks for the reply. At this point the cluster recovered, so I am no longer in that situation. I'll try to go back, reproduce and post the pg query for one of those degraded PGs later on. Cheers Goncalo On 08/27/2015 10:02 PM, Gregory Farnum wrote: On Thu, Aug 27, 2015 at 2

Re: [ceph-users] Rados: Undefined symbol error

2015-08-27 Thread Brad Hubbard
- Original Message - > From: "Aakanksha Pudipeddi-SSI" > To: "Brad Hubbard" > Cc: "Jason Dillaman" , ceph-us...@ceph.com > Sent: Friday, 28 August, 2015 6:15:12 AM > Subject: RE: [ceph-users] Rados: Undefined symbol error > > Hello Brad, > > Thank you for your response. Looks like the c

[ceph-users] 答复: shutdown primary monitor, status of the osds on it will not change, and command like 'rbd create xxx' would block

2015-08-27 Thread zhao.ming...@h3c.com
ceph version:0.94.2 ceph.conf: [global] fsid = 51b9437c-b49d-460c-b919-4bd86cc4d97d mon_initial_members = ubuntu2, ubuntu3, ubuntu4 mon_host = 192.168.33.182,192.168.33.183,192.168.33.184 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_o

Re: [ceph-users] RAM usage only very slowly decreases after cluster recovery

2015-08-27 Thread Haomai Wang
Yes, we already notice this, and have PR to fix partial of this I think https://github.com/ceph/ceph/pull/5451/files On Fri, Aug 28, 2015 at 4:59 AM, Chad William Seys wrote: > Hi all, > > It appears that OSD daemons only very slowly free RAM after an extended period > of an unhealthy cluster (sh

Re: [ceph-users] Can't mount Cephfs

2015-08-27 Thread Goncalo Borges
Hey Andrzej... As Jan replied, I would first try to recover what I can from the ceph cluster. For the time being, I would not be concerned with cephfs. I would also backup the current OSDs so that, if something goes wrong, I can go back to the current state. The recover of the cluster would

Re: [ceph-users] RAM usage only very slowly decreases after cluster recovery

2015-08-27 Thread Somnath Roy
Slow memory release could also be because of tcmalloc. Tcmalloc doesn't release the memory the moment application issue a 'delete' but it cached it inside for future use. If it is not a production cluster and you have spare time to reproduce this, I would suggest to build Ceph code with jemalloc

Re: [ceph-users] Rados: Undefined symbol error

2015-08-27 Thread Brad Hubbard
- Original Message - > From: "Brad Hubbard" > To: "Aakanksha Pudipeddi-SSI" > Cc: ceph-us...@ceph.com > Sent: Friday, 28 August, 2015 10:54:04 AM > Subject: Re: [ceph-users] Rados: Undefined symbol error > > - Original Message - > > From: "Aakanksha Pudipeddi-SSI" > > To: "Brad