Re: [ceph-users] Issue with journal on another drive

2015-09-29 Thread Bill Sharer
I think I got over 10% improvement when I changed from cooked journal file on btrfs based system SSD to a raw partition on the system SSD. The cluster I've been testing with is all consumer grade stuff running on top of AMD piledriver and kaveri based mobo's with the on-board SATA. My SSDs ar

[ceph-users] hadoop on cephfs

2016-04-27 Thread Bill Sharer
Just got into a discussion today where I may have a chance to do work with a db guy who wants hadoop and I want to steer him to it on cephfs. While I'd really like to run gentoo with either infernalis or jewel (when it becomes stable in portage), odds are more likely that I will be required to

Re: [ceph-users] hadoop on cephfs

2016-04-29 Thread Bill Sharer
Actually this guy is already a fan of Hadoop. I was just wondering whether anyone has been playing around with it on top of cephfs lately. It seems like the last round of papers were from around cuttlefish. On 04/28/2016 06:21 AM, Oliver Dzombic wrote: Hi, bad idea :-) Its of course nice a

Re: [ceph-users] Deploying ceph by hand: a few omissions

2016-05-01 Thread Bill Sharer
Actually you didn't need to do a udev rule for raw journals. Disk devices in gentoo have their group ownership set to 'disk'. I only needed to drop ceph into that in /etc/group when going from hammer to infernalis. Did you poke around any of the ceph howto's on the gentoo wiki? It's been a

Re: [ceph-users] Deploying ceph by hand: a few omissions

2016-05-01 Thread Bill Sharer
ma upgrade. I think that one got stablized before it was quite ready for prime time. Even then I'll probably take a good long time to backup some stuff before I try out the shiny new fsck utility. On 05/01/2016 07:13 PM, Stuart Longland wrote: Hi Bill, On 02/05/16 04:37, Bill Sharer w

Re: [ceph-users] Disk failures

2016-06-14 Thread Bill Sharer
rot. The only errors I've run into are from hard shutdowns and possible ecc errors due to working with consumer hardware and memory. I've been on top of btrfs using gentoo since Firefly. Bill Sharer On 06/14/2016 09:27 PM, Christian Balzer wrote: Hello, On Tue, 14 Jun 2016 14:2

[ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-03 Thread Bill Sharer
:_ms_dispatch(Message*)+0x581) [0x556dffe22b91] 13: (Monitor::ms_dispatch(Message*)+0x23) [0x556dffe41393] 14: (DispatchQueue::entry()+0x7ba) [0x556e002e722a] 15: (DispatchQueue::DispatchThread::entry()+0xd) [0x556e001d62cd] 16: (()+0x74a4) [0x7f8290f904

Re: [ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-05 Thread Bill Sharer
ptopp -debug -lttng -tcmalloc {-test} -zfs" PYTHON_TARGETS="python2_7 python3_4 -python3_5" 11,271 KiB Bill Sharer On 07/05/2016 01:45 PM, Gregory Farnum wrote: Thanks for the report; created a ticket and somebody will get on it shortly. http://tracker.ceph.com/issues/16592 -Greg

Re: [ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-06 Thread Bill Sharer
I noticed on that USE list that the 10.2.2 ebuild introduced a new cephfs emerge flag, so I enabled that and emerged everywhere again. The active mon is still crashing on the assertion though. Bill Sharer On 07/05/2016 08:14 PM, Bill Sharer wrote: Relevant USE flags FWIW # emerge -pv

Re: [ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-06 Thread Bill Sharer
Manual downgrade to 10.2.0 put me back in business. I'm going to mask 10.2.2 and then try to let 10.2.1 emerge. Bill Sharer On 07/06/2016 02:16 PM, Bill Sharer wrote: I noticed on that USE list that the 10.2.2 ebuild introduced a new cephfs emerge flag, so I enabled that and em

Re: [ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-08 Thread Bill Sharer
have a problem though if I do a complete cold start of the cluster. Bill Sharer On 07/06/2016 04:19 PM, Bill Sharer wrote: Manual downgrade to 10.2.0 put me back in business. I'm going to mask 10.2.2 and then try to let 10.2.1 emerge. Bill Sharer On 07/06/2016 02:16 PM, Bill Sharer

Re: [ceph-users] ONE pg deep-scrub blocks cluster

2016-07-28 Thread Bill Sharer
ok at the SMART info on the drive or drives in the RAID set to see what the error counts suggest about this. You may also be looking at a drive that's about to fail. Bill Sharer On 07/28/2016 08:46 AM, c wrote: Hello Ceph alikes :) i have a strange issue with one PG (0.223) combined

Re: [ceph-users] ONE pg deep-scrub blocks cluster

2016-07-28 Thread Bill Sharer
hint may be buried in the raw read error rate, seek error rate or other error counts like ecc or crc errors. The long test you are running may or may not show any new information. Bill Sharer On 07/28/2016 11:46 AM, c wrote: Am 2016-07-28 15:26, schrieb Bill Sharer: I suspect the data for

Re: [ceph-users] what happen to the OSDs if the OS disk dies?

2016-08-12 Thread Bill Sharer
keep in mind that you are down 11 osds while that system drive gets rebuilt though. It's safer to do 10 osds and then have a mirror set for the system disk. Bill Sharer On 08/12/2016 03:33 PM, Ronny Aasen wrote: On 12.08.2016 13:41, Félix Barbeira wrote: Hi, I'm planning to m

[ceph-users] assertion error trying to start mds server

2017-10-10 Thread Bill Sharer
I've been in the process of updating my gentoo based cluster both with new hardware and a somewhat postponed update.  This includes some major stuff including the switch from gcc 4.x to 5.4.0 on existing hardware and using gcc 6.4.0 to make better use of AMD Ryzen on the new hardware.  The existing

Re: [ceph-users] assertion error trying to start mds server

2017-10-11 Thread Bill Sharer
07 AM, John Spray wrote: > On Wed, Oct 11, 2017 at 1:42 AM, Bill Sharer wrote: >> I've been in the process of updating my gentoo based cluster both with >> new hardware and a somewhat postponed update. This includes some major >> stuff including the switch from gcc 4.x to 5.

Re: [ceph-users] assertion error trying to start mds server

2017-10-12 Thread Bill Sharer
bluestore.  So far so good...  I just took a peek and saw the files being owned by Mr root though.  Is there going to be an ownership reset at some point or will I have to resolve that by hand? On 10/12/2017 06:09 AM, John Spray wrote: > On Thu, Oct 12, 2017 at 12:23 AM, Bill Sharer wrote: &

[ceph-users] Infernalis OSD errored out on journal permissions without mentioning anything in its log

2016-03-31 Thread Bill Sharer
al process would exit without writing anything to its /var/log/ceph/ceph-osd.n.log. I would have thought there might have been some sort of permission error logged, but nope :-) Bill Sharer ___ ceph-users mailing list ceph-users@lists.cep

Re: [ceph-users] clock skew

2019-04-25 Thread Bill Sharer
hardware newer than about five years old, you might want to look at enabling the NIC clocks using LinuxPTP to keep clock jitter down inside your LAN.  I wrote this article on the Gentoo wiki on enabling PTP in chrony. https://wiki.gentoo.org/wiki/Chrony_with_hardware_timestamping Bill Sharer

Re: [ceph-users] Ceph + SAMBA (vfs_ceph)

2019-08-28 Thread Bill Sharer
igure out what sort of insanity is going on in your IT department.  If you aren't the domain admin or aren't good friends with one who also knows Unix/Linux you may never get anywhere. Bill Sharer On 8/28/19 2:32 PM, Salsa wrote: This is the result: # testparm -s Load smb config