[ceph-users] PG Recovery: HEALTH_ERR to HEALTH_OK

2014-06-03 Thread Jason Harley
Howdy — I’ve had a failure on a small, Dumpling (0.67.4) cluster running on Ubuntu 13.10 machines. I had three OSD nodes (running 6 OSDs each), and lost two of them in a beautiful failure. One of these nodes even went so far as to scramble the XFS filesystems of my OSD disks (I’m curious if i

Re: [ceph-users] PG Recovery: HEALTH_ERR to HEALTH_OK

2014-06-03 Thread Jason Harley
1", > "objects": []}, > "peer_backfill_info": { "begin": "0\/\/0\/\/-1", > "end": "0\/\/0\/\/-1", > "objects": []}, > "ba

Re: [ceph-users] PG Recovery: HEALTH_ERR to HEALTH_OK

2014-06-03 Thread Jason Harley
On Jun 3, 2014, at 5:58 PM, Smart Weblications GmbH - Florian Wiessner wrote: > I think it would be less painfull if you had removed and the immediatelly > recreate the corrupted osd again to avoid 'holes' in the osd ids. It should > work > with your configuration anyhow, though. I agree with

Re: [ceph-users] PG Recovery: HEALTH_ERR to HEALTH_OK

2014-06-05 Thread Jason Harley
, Jason Harley wrote: > On Jun 3, 2014, at 5:58 PM, Smart Weblications GmbH - Florian Wiessner > wrote: > >> I think it would be less painfull if you had removed and the immediatelly >> recreate the corrupted osd again to avoid 'holes' in the osd ids. It should >

[ceph-users] REST API and uWSGI?

2014-06-16 Thread Jason Harley
Howdy — I’d like to run the ceph REST API behind nginx, and uWSGI and UNIX sockets seems like a smart way to do this. Has anyone attempted to get this setup working? I’ve tried writing a uWSGI wrapper as well as just telling ‘uwsgi’ to call the ‘ceph_rest_api’ module without luck. ./JRH

Re: [ceph-users] REST API and uWSGI?

2014-06-17 Thread Jason Harley
On Jun 16, 2014, at 8:52 PM, Wido den Hollander wrote: >> Op 16 jun. 2014 om 19:23 heeft "Jason Harley" het >> volgende geschreven: >> >> Howdy — >> >> I’d like to run the ceph REST API behind nginx, and uWSGI and UNIX sockets >> seems l

[ceph-users] mon: leveldb checksum mismatch

2014-07-03 Thread Jason Harley
Hi list — I’ve got a small dev. cluster: 3 OSD nodes with 6 disks/OSDs each and a single monitor (this, it seems, was my mistake). The monitor node went down hard and it looks like the monitor’s db is in a funny state. Running ‘ceph-mon’ manually with ‘debug_mon 20’ and ‘debug_ms 20’ gave the

Re: [ceph-users] mon: leveldb checksum mismatch

2014-07-03 Thread Jason Harley
Hi Joao, On Jul 3, 2014, at 7:57 PM, Joao Eduardo Luis wrote: > We don't have a way to repair leveldb. Having multiple monitors usually help > with such tricky situations. I know this, but for this small dev cluster I wasn’t thinking about corruption of my mon’s backing store. Silly me :)

Re: [ceph-users] Managing OSDs on twin machines

2014-08-18 Thread Jason Harley
Hi Pierre — You can manipulate your CRUSH map to make use of ‘chassis’ in addition to the default ‘host’ type. I’ve done this with FatTwin and FatTwin^2 boxes with great success. For more reading take a look at: http://ceph.com/docs/master/rados/operations/crush-map/ In particular the ‘Move