Re: [ceph-users] pg_num not being set to ceph.conf default when creating pool via python librados

2015-01-26 Thread Gregory Farnum
Just from memory, I think these values are only used for the initial pool creations when the cluster is first set up. We have been moving for a while to making users specify pg_num explicitly on every pool create, and you should do so. :) -Greg On Mon, Jan 26, 2015 at 7:38 AM Jason Anderson < jaso

Re: [ceph-users] pg_num not being set to ceph.conf default when creating pool via python librados

2015-01-26 Thread Gregory Farnum
t; is out of date? create_pool in rados.py > (https://github.com/ceph/ceph/blob/master/src/pybind/rados.py#L535) only > requires a pool_name….doesn’t even offer pg_num as an optional argument. > > > > Thank you, > > -Jason > > > > From: Gregory Farnum [mailto:g...@gregs4

Re: [ceph-users] osd crush create-or-move doesn't move things?

2015-01-26 Thread Gregory Farnum
On Mon, Jan 26, 2015 at 2:13 PM, Brian Rak wrote: > I have an existing cluster where all the hosts were just added directly, for > example: > > # ceph osd tree > # idweight type name up/down reweight > -1 60.06 root default > ... > -14 1.82host OSD75 > 12 1.8

Re: [ceph-users] Appending to a rados object with feedback

2015-01-26 Thread Gregory Farnum
On Mon, Jan 26, 2015 at 6:47 PM, Kim Vandry wrote: > Hello Ceph users, > > In our application, we found that we have a use case for appending to a > rados object in such a way that the client knows afterwards at what offset > the append happened, even while there may be other concurrent clients do

Re: [ceph-users] chattr +i not working with cephfs

2015-01-28 Thread Gregory Farnum
On Wed, Jan 28, 2015 at 5:24 AM, John Spray wrote: > We don't implement the GETFLAGS and SETFLAGS ioctls used for +i. > > Adding the ioctls is pretty easy, but then we need somewhere to put > the flags. Currently we don't store a "flags" attribute on inodes, > but maybe we could borrow the high b

Re: [ceph-users] cephfs modification time

2015-01-28 Thread Gregory Farnum
Armstrong wrote: > Hey folks, > > Any update on this fix getting merged? We suspect other crashes based on > this bug. > > Thanks, > > Chris > > On Tue, Jan 13, 2015 at 7:09 AM, Gregory Farnum wrote: >> >> Awesome, thanks for the bug report and the fix, guys

Re: [ceph-users] chattr +i not working with cephfs

2015-01-28 Thread Gregory Farnum
On Wed, Jan 28, 2015 at 10:06 AM, Sage Weil wrote: > On Wed, 28 Jan 2015, John Spray wrote: >> On Wed, Jan 28, 2015 at 5:23 PM, Gregory Farnum wrote: >> > My concern is whether we as the FS are responsible for doing anything >> > more than storing and returning t

Re: [ceph-users] cephfs - disabling cache on client and on OSDs

2015-01-30 Thread Gregory Farnum
I don't think there's any way to force the OSDs to do that. What exactly are you trying to do? -Greg On Fri, Jan 30, 2015 at 4:02 AM, Mudit Verma wrote: > Hi All, > > We are working on a project where we are planning to use Ceph as storage. > However, for one experiment we are required to disable

Re: [ceph-users] Blocked requests during and after CephFS delete

2013-12-08 Thread Gregory Farnum
On Sun, Dec 8, 2013 at 7:16 AM, Oliver Schulz wrote: > Hello Ceph-Gurus, > > a short while ago I reported some trouble we had with our cluster > suddenly going into a state of "blocked requests". > > We did a few tests, and we can reproduce the problem: > During / after deleting of a substantial c

Re: [ceph-users] how to set up disks in the same host

2013-12-09 Thread Gregory Farnum
On Mon, Dec 9, 2013 at 1:17 AM, Robert van Leeuwen wrote: >> your client writes the file to one osd, and before this osd acknowledges >> your write request, >> it ensure that it is copied to other osd(s). > > I think this behaviour depends on how you configure you POOL: > > osd pool default min s

Re: [ceph-users] Mount error 12 = Cannot allocate memory

2013-12-09 Thread Gregory Farnum
On Wed, Dec 4, 2013 at 7:15 AM, Mr.Salvatore Rapisarda wrote: > Hi, > > i have a ceph cluster with 3 nodes on Ubuntu 12.04.3 LTS and ceph version > 0.72.1 > > My configuration is the follow: > > * 3 MON > - XRVCLNOSTK001=10.170.0.110 > - XRVCLNOSTK002=10.170.0.111 > - XRVOSTKMNG001=10.170.0.

Re: [ceph-users] Blocked requests during and after CephFS delete

2013-12-09 Thread Gregory Farnum
[ Re-added the list since I don't have log files. ;) ] On Mon, Dec 9, 2013 at 5:52 AM, Oliver Schulz wrote: > Hi Greg, > > I'll send this privately, maybe better not to post log-files, etc. > to the list. :-) > > >> Nobody's reported it before, but I think the CephFS MDS is sending out >> too man

Re: [ceph-users] cache pool/ crush map

2013-12-13 Thread Gregory Farnum
On Wed, Dec 11, 2013 at 6:13 PM, Sherry Shahbazi wrote: > > Hi all, > > I was wondering if u could answer my question regarding cache pool: > If I have got two servers with 1 SSD in front of each of them, what CRUSH > map would be like? > > For example: > If I have defined the following CRUSH map:

Re: [ceph-users] ulimit max user processes (-u) and non-root ceph clients

2013-12-16 Thread Gregory Farnum
On Mon, Dec 16, 2013 at 11:08 AM, Dan van der Ster wrote: > Hi, > > Sorry to revive this old thread, but I wanted to update you on the current > pains we're going through related to clients' nproc (and now nofile) > ulimits. When I started this thread we were using RBD for Glance images > only, bu

Re: [ceph-users] USB pendrive as boot disk

2013-12-16 Thread Gregory Farnum
On Mon, Dec 16, 2013 at 4:35 AM, Gandalf Corvotempesta wrote: > 2013/11/7 Kyle Bader : >> Ceph handles it's own logs vs using syslog so I think your going to have to >> write to tmpfs and have a logger ship it somewhere else quickly. I have a >> feeling Ceph logs will eat a USB device alive, espec

Re: [ceph-users] Ceph incomplete pg

2013-12-18 Thread Gregory Farnum
On Thu, Dec 12, 2013 at 10:58 PM, Jeppesen, Nelson wrote: > I have an issue with incomplete pgs, I’ve tried repairing it but no such > luck. Any ideas what to check? Have you looked at http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg? In particular, what's the output of "ceph

Re: [ceph-users] Failure probability with largish deployments

2013-12-19 Thread Gregory Farnum
On Thu, Dec 19, 2013 at 12:39 AM, Christian Balzer wrote: > > Hello, > > In my "Sanity check" thread I postulated yesterday that to get the same > redundancy and resilience for disk failures (excluding other factors) as > my proposed setup (2 nodes, 2x 11 3TB HDs RAID6 per node, 2 > global hotspar

Re: [ceph-users] ceph-deploy issues with initial mons that aren't up

2013-12-19 Thread Gregory Farnum
"mon initial members" is a race prevention mechanism whose purpose is to prevent your monitors from forming separate quorums when they're brought up by automated software provisioning systems (by not allowing monitors to form a quorum unless everybody in the list is a member). If you want to add ot

Re: [ceph-users] ceph-deploy issues with initial mons that aren't up

2013-12-20 Thread Gregory Farnum
t;> -----Original Message- >> From: Gregory Farnum [mailto:g...@inktank.com] >> Sent: Thursday, December 19, 2013 10:19 PM >> To: Don Talton (dotalton) >> Cc: ceph-users@lists.ceph.com >> Subject: Re: [ceph-users] ceph-deploy issues with initial mons that are

Re: [ceph-users] public network

2014-01-05 Thread Gregory Farnum
They need to be able to send messages to each other, but routing can work however you like. On Sunday, January 5, 2014, James Harper wrote: > Is there any requirement that the monitors have to be on the same subnet > as each other, and/or the osd public network? It's going to simplify things > gr

Re: [ceph-users] [Rados] How to get the scrub progressing ?

2014-01-06 Thread Gregory Farnum
On Mon, Dec 30, 2013 at 11:14 PM, Kuo Hugo wrote: > > Hi all, > > I have several question about osd scrub. > > Does the scrub job run in the background automatically? Is it working > periodically ? Yes, the OSDs will periodically scrub the PGs they host based on load and the min/max scrub interv

Re: [ceph-users] ceph osd perf question

2014-01-06 Thread Gregory Farnum
On Fri, Jan 3, 2014 at 2:02 AM, Andrei Mikhailovsky wrote: > Hi guys, > > Could someone explain what's the new perf stats show and if the numbers are > reasonable on my cluster? > > I am concerned about the high fs_commit_latency, which seems to be above > 150ms for all osds. I've tried to find th

Re: [ceph-users] Ceph Command Prepending "None" to output on one node (only)

2014-01-06 Thread Gregory Farnum
I have a vague memory of this being something that happened in an outdated version of the ceph tool. Are you running an older binary on the node in question? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sat, Jan 4, 2014 at 4:34 PM, Zeb Palmer wrote: > I have a small ceph

Re: [ceph-users] CephFS files not appearing in DF (or rados ls)

2014-01-06 Thread Gregory Farnum
On Thu, Jan 2, 2014 at 2:18 PM, Alex Pearson wrote: > Hi All, > Victory! Found the issue, it was a mistake on my part, however it does raise > another questions... > > The issue was: > root@osh1:~# ceph --cluster apics auth list > installed auth entries: > > client.cuckoo > key: AQBjTbl

Re: [ceph-users] How can I set the warning level?

2014-01-06 Thread Gregory Farnum
On Wed, Dec 25, 2013 at 6:13 PM, vernon1...@126.com wrote: > Hello, my Mon's always HEALTH_WARN, and I run ceph health detail, it show > me like this: > > HEALTH_WARN > mon.2 addr 192.168.0.7:6789/0 has 30% avail disk space -- low disk space! > > I want to know how to set this warning level? I ha

Re: [ceph-users] cannot see recovery statistics + pgs stuck unclean

2014-01-06 Thread Gregory Farnum
[Hrm, this email was in my spam folder.] At a quick glance, you're probably running into some issues because you've got two racks of very different weights. Things will probably get better if you enable the optimal "crush tunables"; check out the docs on that and see if you can switch to them. -Gr

Re: [ceph-users] What's the status of feature: S3 object versioning?

2014-01-06 Thread Gregory Farnum
On Thu, Jan 2, 2014 at 12:40 AM, Ray Lv wrote: > Hi there, > > Noted that there is a Blueprint item about S3 object versioning in radosgw > for Firefly at > http://wiki.ceph.com/Planning/Blueprints/Firefly/rgw%3A_object_versioning > And Sage has announced v0.74 release for Firefly. Do you guys kno

Re: [ceph-users] 4 PGs stuck inactive

2014-01-07 Thread Gregory Farnum
Assuming the one who lost its filesystem is totally gone, mark it lost. That will tell the OSDs to give up on whatever data it might have had and you should be good to go (modulo whatever data you might have lost from only having it on the dead OSD during the reboot). -Greg Software Engineer #42 @

Re: [ceph-users] 4 PGs stuck inactive

2014-01-07 Thread Gregory Farnum
Oh, sorry, you did do that. Hrm. What osdmap epoch did your lost node (0, I assume) disappear in? What version of Ceph are you running? That pg stat isn't making a lot of sense to me. Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, Jan 7, 2014 at 2:45 PM, Gregory F

Re: [ceph-users] backfilling after OSD marked out _and_ OSD removed

2014-01-09 Thread Gregory Farnum
On Thu, Jan 9, 2014 at 6:27 AM, Dan Van Der Ster wrote: > Here’s a more direct question. Given this osd tree: > > # ceph osd tree |head > # idweight type name up/down reweight > -1 2952root default > -2 2952room 0513-R-0050 > -3 262.1 ra

Re: [ceph-users] backfilling after OSD marked out _and_ OSD removed

2014-01-09 Thread Gregory Farnum
Yep! On Thu, Jan 9, 2014 at 11:01 AM, Dan Van Der Ster wrote: > Thanks Greg. One thought I had is that I might try just crush rm'ing the OSD > instead of or just after marking it out... That should avoid the double > rebalance, right? > > Cheers, Dan > > On Jan 9, 20

Re: [ceph-users] RBD cache questions (kernel vs. user space, KVM live migration)

2014-01-14 Thread Gregory Farnum
On Tuesday, January 14, 2014, Christian Balzer wrote: > > Hello, > > In http://ceph.com/docs/next/rbd/rbd-config-ref/ it is said that: > > "The kernel driver for Ceph block devices can use the Linux page cache to > improve performance." > > Is there anywhere that provides more details about this?

Re: [ceph-users] One OSD always dieing

2014-01-15 Thread Gregory Farnum
Hrm, at first glance that looks like the on-disk state got corrupted somehow. If it's only one OSD which has this issue, I'd turn it off and mark it out. Then if the cluster recovers properly, wipe it and put it back in as a new OSD. -Greg On Wed, Jan 15, 2014 at 1:49 AM, Rottmann, Jonas (centron

Re: [ceph-users] cepfs: Minimal deployment

2014-01-17 Thread Gregory Farnum
On Friday, January 17, 2014, Iban Cabrillo wrote: > Dear, > we are studying the possibility to migrate our FS in the next year to > cephfs. I know that it is not prepare for production environments yet, but > we are planning to play with it in the next months deploying a basic > testbed. > Re

Re: [ceph-users] how does ceph handle object writes?

2014-01-21 Thread Gregory Farnum
Almost! The primary OSD sends out the data to its replicas simultaneously with putting it into the journal. -Greg On Monday, January 20, 2014, Tim Zhang wrote: > Hi guys, > I wonder how does store objects. Consider the writing obj process, IMO, > osd first get obj data from client, then the pri

Re: [ceph-users] I/O Performance Statistics

2014-01-22 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 8:04 AM, Dan Ryder (daryder) wrote: > > Hello, > > > > I am wondering if there is any detailed documentation for obtaining I/O > statistics for a Ceph cluster. > > The important metrics I’m looking for are: the number of operations, size of > operations, and latency of op

Re: [ceph-users] placing SSDs and SATAs pool in same hosts

2014-01-22 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 8:35 AM, zorg wrote: > Hi, > After reading the thread > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-June/002358.html > > We have done this crush map to make thing work. > srv1 and srv1ssd are the same physical server (same srv2,3,4) > we split it in the crush t

Re: [ceph-users] Reducing I/O when increasing number of PGs

2014-01-22 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 9:13 AM, Caius Howcroft wrote: > Hi all, > > > I want to double the number of pgs available for a pool, however I > want to reduce as much as possible the resulting I/O storm (I have > quite a bit of data in these pools). > > What is the best way of doing this? Is it using

Re: [ceph-users] Problem with Mounting to Pool 0 with CephFS

2014-01-22 Thread Gregory Farnum
On Sun, Jan 19, 2014 at 9:00 PM, Sherry Shahbazi wrote: > Hi all, > > I have three pools, which I want to mount Pool 0 with CephFS. > When I try to set the layout by changing the pool to to 0 (cephfs > /mnt/oruafs/pool0/ set_layout -p 0), it would not be set to pool 0 while I > am able to set it t

Re: [ceph-users] how does ceph handle object writes?

2014-01-22 Thread Gregory Farnum
[Re-added the list.] On Wed, Jan 22, 2014 at 4:15 AM, Tim Zhang wrote: > Hi Gregory, > I find in the src code:ReplicatedPG.cc in the function do_pg_op() I see this > two flag: > CEPH_OSD_FLAG_ACK and CEPH_OSD_FLAG_ONDISK, > what's the differenrence between this two flag? In some circumstances (w

Re: [ceph-users] OSD port usage

2014-01-22 Thread Gregory Farnum
On Tue, Jan 21, 2014 at 8:26 AM, Sylvain Munaut wrote: > Hi, > > I noticed in the documentation that the OSD should use 3 ports per OSD > daemon running and so when I setup the cluster, I originally opened > enough port to accomodate this (with a small margin so that restart > could proceed even i

Re: [ceph-users] failed to open snapshot after 'rados cppool '

2014-01-22 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 3:23 AM, Dmitry Lysenko wrote: > Good day. > > Some time ago i change pg_num like this > http://www.sebastien-han.fr/blog/2013/03/12/ceph-change-pg-number-on-the-fly/: > > ceph osd pool create one-new 500 > rados cppool one one-new Unfortunately, this command is not copyi

Re: [ceph-users] Reducing I/O when increasing number of PGs

2014-01-22 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 3:50 PM, bf wrote: > > > Gregory Farnum writes: > >> >> On Wed, Jan 22, 2014 at 9:13 AM, Caius Howcroft >> > I want to double the number of pgs available for a pool, however I >> > want to reduce as much as possible the resulting

Re: [ceph-users] Mounting ceph-fuse with fstab

2014-01-23 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 3:23 PM, Karol Kozubal wrote: > Hi Everyone, > > I have a few questions concerning mounting cephfs with ceph-fuse in fstab at > boot. I am currently successfully mounting cephfs using ceph-fuse on 6 > clients. I use the following command, where the ip is my mon address: >

Re: [ceph-users] Usually Offline OSDi

2014-01-23 Thread Gregory Farnum
On Wed, Jan 22, 2014 at 10:43 PM, Schlacta, Christ wrote: > can ceph handle a configuration where a custer node is not "always on", but > rather gets booted periodically to sync to the cluster, and is also > sometimes up full time as demand requires? I ask because I want to put an > OSD on each o

Re: [ceph-users] Reducing I/O when increasing number of PGs

2014-01-23 Thread Gregory Farnum
On Thu, Jan 23, 2014 at 3:35 AM, bf wrote: > Gregory Farnum writes: > > >> Yes, Ceph does all the heavy lifting. Multiple PGs with the same OSDs >> can happen (eg, if you only have two OSDs, all PGs will be on both), >> but it behaves about as well as is possible with

Re: [ceph-users] failed to open snapshot after 'rados cppool '

2014-01-23 Thread Gregory Farnum
[ Returning list to thread. ] On Wed, Jan 22, 2014 at 11:37 PM, Dmitry Lysenko wrote: > 22.01.2014 13:01, Gregory Farnum пишет: > > >> On Wed, Jan 22, 2014 at 3:23 AM, Dmitry Lysenko wrote: >> > Good day. >> > >> > Some time ago i change pg_num like t

Re: [ceph-users] Mounting ceph-fuse with fstab

2014-01-23 Thread Gregory Farnum
On Thu, Jan 23, 2014 at 8:07 AM, Arne Wiebalck wrote: > > On Jan 23, 2014, at 4:18 PM, Gregory Farnum > wrote: > >> On Wed, Jan 22, 2014 at 3:23 PM, Karol Kozubal >> wrote: >>> Hi Everyone, >>> >>> I have a few questions concerning mounti

Re: [ceph-users] Ceph vs zfs data integrity

2014-01-23 Thread Gregory Farnum
On Thu, Jan 23, 2014 at 2:21 PM, Schlacta, Christ wrote: > What guarantees does ceph place on data integrity? Zfs uses a Merkel tree to > guarantee the integrity of all data and metadata on disk and will ultimately > refuse to return "duff" data to an end user consumer. > > I know ceph provides so

Re: [ceph-users] OSD/monitor timeouts?

2014-01-23 Thread Gregory Farnum
On Thu, Jan 23, 2014 at 5:24 PM, Stuart Longland wrote: > Hi all, > > I'm in the process of setting up a storage cluster for production use. > At the moment I have it in development and am testing the robustness of > the cluster. One key thing I'm conscious of is single points of > failure. Thus

Re: [ceph-users] OSD port usage

2014-01-24 Thread Gregory Farnum
're giving blood.” > > Phone: +33 (0)1 49 70 99 72 > Mail: sebastien@enovance.com > Address : 10, rue de la Victoire - 75009 Paris > Web : www.enovance.com - Twitter : @enovance > > On 22 Jan 2014, at 19:01, Gregory Farnum > > wrote: > > > On Tue, Jan 21

Re: [ceph-users] many meta files in osd

2014-01-27 Thread Gregory Farnum
Looks like you got lost over the Christmas holidays; sorry! I'm not an expert on running rgw but it sounds like garbage collection isn't running or something. What version are you on, and have you done anything to set it up? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Su

Re: [ceph-users] About Ceph Performance Counters

2014-01-27 Thread Gregory Farnum
These aren't really consumerized yet, so you pretty much have to google and see if somebody's already discussed them or go through the code. Not sure where they are on the priority list for docs. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sun, Jan 26, 2014 at 7:34 PM, D

Re: [ceph-users] booting from a CEPH rbd or CEPH filesystem

2014-01-27 Thread Gregory Farnum
This isn't a topic I know a ton about, but: It is not possible to boot from CephFS, but will be soon (search for "[PATCH 1/4] init: Add a new root device option, the Ceph file system"). I think it is possible to boot from rbd (there is native kernel support for it as a block device, for starters),

Re: [ceph-users] OSD/monitor timeouts?

2014-01-27 Thread Gregory Farnum
On Mon, Jan 27, 2014 at 9:05 PM, Stuart Longland wrote: > On 25/01/14 16:41, Stuart Longland wrote: >> Hi Gregory, >> On 24/01/14 12:20, Gregory Farnum wrote: >>> Did the cluster actually detect the node as down? (You could check >>> this by looking at the ceph

Re: [ceph-users] OSD/monitor timeouts?

2014-01-29 Thread Gregory Farnum
On Tue, Jan 28, 2014 at 6:43 PM, Stuart Longland wrote: > Hi Gregory, > On 28/01/14 15:51, Gregory Farnum wrote: >>> I do note ntp doesn't seem to be doing its job, but that's a side issue. >> Actually, that could be it. If you take down one of the monitors and &g

Re: [ceph-users] RGW Replication

2014-02-03 Thread Gregory Farnum
On Mon, Feb 3, 2014 at 10:43 AM, Craig Lewis wrote: > I've been noticing somethings strange with my RGW federation. I added some > statistics to radosgw-agent to try and get some insight > (https://github.com/ceph/radosgw-agent/pull/7), but that just showed me that > I don't understand how replic

Re: [ceph-users] Low RBD Performance

2014-02-04 Thread Gregory Farnum
On Tue, Feb 4, 2014 at 9:29 AM, Gruher, Joseph R wrote: > > >>-Original Message- >>From: ceph-users-boun...@lists.ceph.com [mailto:ceph-users- >>boun...@lists.ceph.com] On Behalf Of Mark Nelson >>Sent: Monday, February 03, 2014 6:48 PM >>To: ceph-users@lists.ceph.com >>Subject: Re: [ceph-u

Re: [ceph-users] Usage of ceph-syn

2014-02-05 Thread Gregory Farnum
ceph-syn is a synthetic client against the full CephFS distributed filesystem. It's really more of a developer tool than even a usable benchmark at this time. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Wed, Feb 5, 2014 at 10:39 AM, Sahana wrote: > Hi , > > I have few q

Re: [ceph-users] How client choose among replications?

2014-02-11 Thread Gregory Farnum
There are a few more options that are in current dev releases for directing traffic to replicas, but it remains pretty specialized and probably won't be supported past the direct librados client layer for Firefly (unless somebody's prioritized it for RGW or RBD that I haven't heard about). -Greg So

Re: [ceph-users] Use of daily-created/-deleted pools

2014-02-12 Thread Gregory Farnum
On Tuesday, February 11, 2014, Hyangtack Lee wrote: > I'm new to Ceph, and looking for a new storage to replace legacy system. > > My system has a lot of files accessing temporarily for 2 or 3 days. > Those files are uploaded from many clients everyday, and batch job deletes > unused files everyd

Re: [ceph-users] Unique identifier in osd debug log to track ONE process from beginning to end

2014-02-12 Thread Gregory Farnum
This is harder than it sounds, but you can get a broad picture by looking for all the instances of the client request string (I believe it's the one from the incoming message, but I'm not at a computer to check). It outputs all the key points with that tag, and you can look at what's going on in th

Re: [ceph-users] Radosgw / Chunked transfer / RHEL / Swift

2014-02-12 Thread Gregory Farnum
If you look at http://ceph.com/docs/master/install/install-ceph-gateway, the fcgi package you have installed is labeled as "W/Out 100 Continue". You need to add a separate "gitbuilder" repository if you want it with 100-continue (supporting chunked transfers); it's all described on that page. -Greg

Re: [ceph-users] How to fix an incomplete PG on an 2 copy ceph-cluster?

2014-02-16 Thread Gregory Farnum
Check out http://ceph.com/docs/master/rados/operations/placement-groups/#get-statistics-for-stuck-pgs and http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg/. What does the dump of the PG say is going on? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sun

Re: [ceph-users] Sudden RADOS Gateway issues caused by missing xattrs

2014-02-16 Thread Gregory Farnum
Did you maybe upgrade that box to v0.67.6? This sounds like one of the bugs Sage mentioned in it. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sun, Feb 16, 2014 at 4:23 AM, Wido den Hollander wrote: > Hi, > > Yesterday I got a notification that a RGW setup was having iss

Re: [ceph-users] Unique identifier in osd debug log to track ONE process from beginning to end

2014-02-18 Thread Gregory Farnum
e able to see the big points of progress. But really, to read these logs you need to spend the time to understand them; there's no quick fix for it. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com > > Thank you, > Mihaly > > 2014-02-12 17:28 GMT+01:00 Gregor

Re: [ceph-users] osd migration question

2014-02-19 Thread Gregory Farnum
On Wed, Feb 19, 2014 at 1:31 PM, mike smith wrote: > I am trying to learn about Ceph and have been looking at the documentation > and speaking to colleagues who work with it and had a question that I could > not get the answer to. As I understand it, the Crush map is updated every > time a disk i

Re: [ceph-users] "Requests are blocked > 32 sec"

2014-02-19 Thread Gregory Farnum
On Wed, Feb 19, 2014 at 2:50 AM, Dane Elwell wrote: > Hello list, > > We've recently moved our 588TB Ceph cluster into production by moving > VMs on there, but this morning we started receiving the following > message: > > cluster etc > health HEALTH_WARN 20 requests are blocked >

Re: [ceph-users] CephFS and slow requests

2014-02-20 Thread Gregory Farnum
Arne, Sorry this got dropped -- I had it marked in my mail but didn't have the chance to think about it seriously when you sent it. Does this still happen after the updatedb config change you guys made recently? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Fri, Jan 31, 20

Re: [ceph-users] Ceph GET latency

2014-02-20 Thread Gregory Farnum
On Tue, Feb 18, 2014 at 7:24 AM, Guang Yang wrote: > Hi ceph-users, > We are using Ceph (radosgw) to store user generated images, as GET latency > is critical for us, most recently I did some investigation over the GET path > to understand where time spend. > > I first confirmed that the latency c

Re: [ceph-users] ceph-mon segmentation fault

2014-02-20 Thread Gregory Farnum
I created a ticket: http://tracker.ceph.com/issues/7487 But my guess is that this is a result of having 0 CRUSH weight for the entire tree while linking them up. Can you give the OSD a weight and see if it works after that? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Tu

Re: [ceph-users] Journal thoughts

2014-02-20 Thread Gregory Farnum
On Mon, Feb 17, 2014 at 12:29 AM, Alex Pearson wrote: > Hi All, > I've been looking, but haven't been able to find any detailed documentation > about the journal usage on OSDs. Does anyone have any detailed docs they > could share? My initial questions are: Hmm, I'm not sure if we have any se

Re: [ceph-users] stuck unclean/stuck inactive

2014-02-20 Thread Gregory Farnum
On Thu, Jan 30, 2014 at 1:17 PM, Derek Yarnell wrote: > Hi, > > So I am trying to remove OSDs from one of our 6 ceph OSDs, this is a > brand new cluster and no data is yet on it. I was following the manual > procedure[1] with the following script. I removed OSDs 0-3 but I am > seeing ceph not fu

Re: [ceph-users] monitor data

2014-02-20 Thread Gregory Farnum
On Thu, Feb 20, 2014 at 7:52 AM, Pavel V. Kaygorodov wrote: > Hi! > > May be it is a dumb question, but anyway: > > If I lose all monitors (mon data dirs), does it possible to recover cluster > with data from OSDs only? Short version: no. Long version: probably, if you're willing to spend lots o

Re: [ceph-users] ceph-mon segmentation fault

2014-02-20 Thread Gregory Farnum
[Re-added the list] On Thu, Feb 20, 2014 at 8:09 AM, Pavel V. Kaygorodov wrote: > Hi! > >> I created a ticket: http://tracker.ceph.com/issues/7487 >> >> But my guess is that this is a result of having 0 CRUSH weight for the >> entire tree while linking them up. Can you give the OSD a weight and >

Re: [ceph-users] Flapping/Crashing OSD

2014-02-20 Thread Gregory Farnum
On Thu, Feb 20, 2014 at 4:26 AM, Michael wrote: > Hi All, > > Have a log full of - > > "log [ERR] : 1.9 log bound mismatch, info (46784'1236417,46797'1239418] > actual [46784'1235968,46797'1239418]" Do you mean that error message is showing up for a lot of different PGs? The specific error indica

Re: [ceph-users] hashpspool and backfilling

2014-02-20 Thread Gregory Farnum
On Tue, Feb 18, 2014 at 8:21 AM, Dan van der Ster wrote: > Hi, > Today I've noticed an interesting result of not have hashpspool > enabled on a number of pools -- backfilling is delayed. > > Take for example the following case: a PG from each of 5 different > pools (details below) are all mapped t

Re: [ceph-users] hashpspool and backfilling

2014-02-20 Thread Gregory Farnum
On Thu, Feb 20, 2014 at 12:48 PM, Dan van der Ster wrote: > Hi, > > On Thu, Feb 20, 2014 at 7:47 PM, Gregory Farnum wrote: >> >> On Tue, Feb 18, 2014 at 8:21 AM, Dan van der Ster >> wrote: >> > Hi, >> > Today I've noticed an interesting result

Re: [ceph-users] How does Ceph deal with OSDs that have been away for a while?

2014-02-21 Thread Gregory Farnum
It depends on how long ago (in terms of data writes) it disappeared. Each PG has a log of the changes that have been made (by default I think it's 3000? Maybe just 1k), and if an OSD goes away and comes back while the logs still overlap it will just sync up the changed objects. Otherwise it has to

Re: [ceph-users] goofy results for df

2014-02-21 Thread Gregory Farnum
I haven't done the math, but it's probably a result of how the df command interprets the output of the statfs syscall. We changed the fr_size and block_size units we report to make it work more consistently across different systems "recently"; I don't know if that change was before or after the ker

Re: [ceph-users] How does Ceph deal with OSDs that have been away for a while?

2014-02-21 Thread Gregory Farnum
in the case where the log no longer overlaps? > > I guess the key question is - do I have to worry about it, or will it > always "do the right thing"? > > Tim. > > On Fri, Feb 21, 2014 at 11:57:09AM -0800, Gregory Farnum wrote: >> It depends on how long ago (in terms

Re: [ceph-users] CephFS and slow requests

2014-02-24 Thread Gregory Farnum
.818 x:6803/128164 1220 : [WRN] slow request >> 30.446683 seconds old, received at 2014-02-21 16:25:44.754914: >> osd_op(client.16735018.1:22524842 10352bf.0355 [write 0~4194304 >> [12@0],startsync 0~0] 0.c36d4557 snapc 1=[] e42655) v4 currently waiting for >> subop

Re: [ceph-users] goofy results for df

2014-02-24 Thread Gregory Farnum
n 20 03:36:48 > UTC > > Markus > Am 21.02.2014 20:59, schrieb Gregory Farnum: > >> I haven't done the math, but it's probably a result of how the df >> command interprets the output of the statfs syscall. We changed the >> fr_size and block_size units

Re: [ceph-users] CephFS and slow requests

2014-02-25 Thread Gregory Farnum
Okay, well, let's try and track some of these down. What's the content of the "ceph.layout" xattr on the directory you're running this test in? Can you verify that pool 0 is the data pool used by CephFS, and that all reported slow ops are in that pool? Can you record the IO patterns on an OSD while

Re: [ceph-users] goofy results for df

2014-02-25 Thread Gregory Farnum
On Mon, Feb 24, 2014 at 11:48 PM, Markus Goldberg wrote: > Hi Gregory, > here we go: > > root@bd-a:/mnt/myceph# > root@bd-a:/mnt/myceph# ls -la > insgesamt 4 > drwxr-xr-x 1 root root 25928099891213 Feb 24 14:14 . > drwxr-xr-x 4 root root 4096 Aug 30 10:34 .. > drwx-- 1 root root 2592

Re: [ceph-users] pages stuck unclean (but remapped)

2014-02-25 Thread Gregory Farnum
With the reweight-by-utilization applied, CRUSH is failing to generate mappings of enough OSDs, so the system is falling back to keeping around copies that already exist, even though they aren't located on the correct CRUSH-mapped OSDs (since there aren't enough OSDs). Are your OSDs correctly weigh

Re: [ceph-users] PG folder hierarchy

2014-02-25 Thread Gregory Farnum
On Tue, Feb 25, 2014 at 7:13 PM, Guang wrote: > Hello, > Most recently when looking at PG's folder splitting, I found that there was > only one sub folder in the top 3 / 4 levels and start having 16 sub folders > starting from level 6, what is the design consideration behind this? > > For example,

Re: [ceph-users] goofy results for df

2014-02-26 Thread Gregory Farnum
osd/ceph-0 > ^^^ wrong ^^ > Am 25.02.2014 18:55, schrieb Gregory Farnum: > > [Re-adding the list.] > > Yeah, that pg dump indicates that each OSD believes it is storing > about 30GB (which could include a lot of stuff besides the raw RADOS > usage) a

Re: [ceph-users] recover from node failure / monitor and osds do not come back

2014-02-26 Thread Gregory Farnum
Your OSDs aren't supposed to be listed in the config file, but they should show up under /var/lib/ceph. Probably your OSD disks aren't being mounted for some reason (that would be the bug). Try mounting them and seeing what blocked the mount. -Greg Software Engineer #42 @ http://inktank.com | http:

Re: [ceph-users] map RBD into CephFS?

2014-02-26 Thread Gregory Farnum
On Wed, Feb 26, 2014 at 6:10 AM, David Champion wrote: > I have a 1.6 TB collecton of 8 million files in CephFS, distributed up > to 8-10 directories deep. (Never mind why - this design decision is out > of my hands and not in scope.) I need to expose this data on multiple > application servers.

Re: [ceph-users] map RBD into CephFS?

2014-02-26 Thread Gregory Farnum
On Wed, Feb 26, 2014 at 10:37 AM, David Champion wrote: > Thanks, Greg, for the response. > > * On 26 Feb 2014, Gregory Farnum wrote: >> > >> > 1. Place the 8m files in a disk image. Mount the disk image (read-only) >> > to provide access to the 8m files,

Re: [ceph-users] Trying to rescue a lost quorum

2014-02-27 Thread Gregory Farnum
On Thu, Feb 27, 2014 at 4:25 PM, Marc wrote: > Hi, > > I was handed a Ceph cluster that had just lost quorum due to 2/3 mons > (b,c) running out of disk space (using up 15GB each). We were trying to > rescue this cluster without service downtime. As such we freed up some > space to keep mon b runn

Re: [ceph-users] Cephfs + CTDB+Samba/NFS

2014-02-28 Thread Gregory Farnum
On Fri, Feb 28, 2014 at 6:14 AM, Sage Weil wrote: > On Wed, 26 Feb 2014, haiquan...@sina.com wrote: >> Hi , >> >> Recently, we are testing use CTDB+Cephfs+Samba/NFS HA cluster , but >> it's not success, so the Cephfs if support CTDB cluster file ? if it >> support , could you pls offer a g

Re: [ceph-users] CephFS: files never stored on OSDs

2014-02-28 Thread Gregory Farnum
By default your filesystem data is stored in the "data" pool, ID 0. You can change to a different pool (for files going forward, not existing ones) by setting the root directory's layout via the ceph.layout.pool virtual xattr, but it doesn't look like you've done that yet. Until then, you've got tw

Re: [ceph-users] Viewing CephFS Client Debugging

2014-02-28 Thread Gregory Farnum
On Thu, Feb 27, 2014 at 9:29 PM, Michael Sevilla wrote: > I'm looking for the debug messages in Client.cc, which uses ldout > (library debugging). I increased the client debug level for all > daemons (i.e. under [global] in ceph.conf) and verified that it got > set: > > $ ceph --admin-daemon /var/

Re: [ceph-users] map RBD into CephFS?

2014-02-28 Thread Gregory Farnum
On Wed, Feb 26, 2014 at 11:39 AM, David Champion wrote: > * On 26 Feb 2014, Gregory Farnum wrote: >> >> > q1. CephFS has a tunable for max file size, currently set to 1TB. If >> >> > I want to change this, what needs to be done or redone? Do I have to >>

Re: [ceph-users] RBD Snapshots

2014-02-28 Thread Gregory Farnum
RBD itself will behave fine with whenever you take the snapshot. The thing to worry about is that it's a snapshot at the block device layer, not the filesystem layer, so if you don't quiesce IO and sync to disk the filesystem might not be entirely happy with you for the same reasons that it won't b

Re: [ceph-users] recover from node failure / monitor and osds do not come back

2014-03-03 Thread Gregory Farnum
On Sunday, March 2, 2014, Diedrich Ehlerding < diedrich.ehlerd...@ts.fujitsu.com> wrote: > Gregory Farnum wrote: > > > > > Your OSDs aren't supposed to be listed in the config file, but they > > should show up under /var/lib/ceph. Probably your OSD disks aren&

Re: [ceph-users] High fs_apply_latency on one node

2014-03-03 Thread Gregory Farnum
The apply latency is how long it's taking for the backing filesystem to ack (not sync to disk) writes from the OSD. Either it's getting a lot more writes than the other OSDs (you can check by seeing how many PGs are mapped to each) and then just apply standard local fs debugging techniques to that

Re: [ceph-users] "full ratio" - how does this work with multiple pools on seprate OSDs?

2014-03-04 Thread Gregory Farnum
The setting is calculated per-OSD, and if any OSD hits the hard limit the whole cluster transitions to the full state and stops accepting writes until the situation is resolved. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, Mar 4, 2014 at 9:58 AM, Barnes, Thomas J wr

<    5   6   7   8   9   10   11   12   13   14   >