Thanks Jay, I'll do some testing with this and report back.

Jason


On Thu, Aug 15, 2013 at 7:10 AM, Jay Kreps <jay.kr...@gmail.com> wrote:

> I believe either should work. The broker has a record of what it should
> have in zk and will recreate any missing logs. Try it to make sure though.
>
> Sent from my iPhone
>
> On Aug 15, 2013, at 12:52 AM, Jason Rosenberg <j...@squareup.com> wrote:
>
> > Ok, that makes sense that the broker will shut itself down.
> >
> > If we bring it back up, can this be with an altered set of log.dirs?
>  Will
> > the destroyed partitions get rebuilt on a new log.dir?  Or do we have to
> > bring it back up with a new or repaired disk, matching the old log.dir,
> in
> > order for those replicas to be rebuilt?
> >
> > Jason
> >
> >
> > On Wed, Aug 14, 2013 at 4:16 PM, Jay Kreps <jay.kr...@gmail.com> wrote:
> >
> >> If you get a disk error that results in an IOException the broker will
> >> shut itself down. You would then have the option of replacing the disk
> or
> >> deleting that data directory from the list. When the broker is brought
> back
> >> up the intact partitions will quickly catch up and be online; the
> destroyed
> >> partitions will have to fully rebuild off the other replicas and will
> take
> >> a little longer but will automatically come back online once they have
> >> restored off the replicas.
> >>
> >> -jay
> >>
> >> Sent from my iPhone
> >>
> >> On Aug 14, 2013, at 1:49 PM, Jason Rosenberg <j...@squareup.com> wrote:
> >>
> >>> I'm getting ready to try out this configuration (use multiple disks, no
> >>> RAID, per broker).  One concern is the procedure for recovering if
> there
> >> is
> >>> a disk failure.
> >>>
> >>> If a disk fails, will the broker go offline, or will it continue
> serving
> >>> partitions on its remaining good disks?  And if so, is there a
> procedure
> >>> for moving the partitions that were on the failed disk, but not
> >> necessarily
> >>> all the others on that broker?
> >>>
> >>> Jason
> >>>
> >>>
> >>> On Thu, Jun 20, 2013 at 3:15 PM, Jason Rosenberg <j...@squareup.com>
> >> wrote:
> >>>
> >>>> yeah, that would work!
> >>>>
> >>>>
> >>>> On Thu, Jun 20, 2013 at 1:20 PM, Jay Kreps <jay.kr...@gmail.com>
> wrote:
> >>>>
> >>>>> Yeah we didn't go as far as adding weighting or anything like that--I
> >>>>> think we'd be open to a patch that did that as long as it was
> >>>>> optional. In the short term you can obviously add multiple
> directories
> >>>>> on the same disk to increase its share.
> >>>>>
> >>>>> -Jay
> >>>>>
> >>>>> On Thu, Jun 20, 2013 at 12:59 PM, Jason Rosenberg <j...@squareup.com>
> >>>>> wrote:
> >>>>>> This sounds like a great idea, to just disks as "just a bunch of
> >> disks"
> >>>>> or
> >>>>>> JBOD.....hdfs works well this way.
> >>>>>>
> >>>>>> Do all the disks need to be the same size, to use them evenly?
>  Since
> >> it
> >>>>>> will allocate partitions randomly?
> >>>>>>
> >>>>>> It would be nice if you had 2 disks, with one twice as large as the
> >>>>> other,
> >>>>>> if the larger would be twice as likely to receive partitions as the
> >>>>> smaller
> >>>>>> one, etc.
> >>>>>>
> >>>>>> I suppose this goes into my earlier question to the list, vis-a-vis
> >>>>>> heterogeneous brokers (e.g. utilize brokers with different sized
> >>>>> storage,
> >>>>>> using some sort of weighting scheme, etc.).
> >>>>>>
> >>>>>> Jason
> >>>>>>
> >>>>>>
> >>>>>> On Thu, Jun 20, 2013 at 11:07 AM, Jay Kreps <jay.kr...@gmail.com>
> >>>>> wrote:
> >>>>>>
> >>>>>>> The intention is to allow the use of multiple disks without RAID or
> >>>>>>> logical volume management. We have found that there are a lot of
> >>>>>>> downsides to RAID--in particular a huge throughput hit. Since we
> >>>>>>> already have a parallelism model due to partitioning and a fault
> >>>>>>> tolerance model with replication RAID doesn't actually buy much.
> With
> >>>>>>> this feature you can directly mount multiple disks as their own
> >>>>>>> directory and the server will randomly assign partitions to them.
> >>>>>>>
> >>>>>>> Obviously this will only work well if there are enough
> >> high-throughput
> >>>>>>> partitions to make load balance evenly (e.g. if you have only one
> big
> >>>>>>> partition per server then this isn't going to work).
> >>>>>>>
> >>>>>>> -Jay
> >>>>>>>
> >>>>>>> On Wed, Jun 19, 2013 at 11:01 PM, Jason Rosenberg <
> j...@squareup.com>
> >>>>>>> wrote:
> >>>>>>>> is it possible for a partition to have multiple replicas on
> >> different
> >>>>>>>> directories on the same broker?  (hopefully no!)
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Wed, Jun 19, 2013 at 10:47 PM, Jun Rao <jun...@gmail.com>
> wrote:
> >>>>>>>>
> >>>>>>>>> It takes a comma separated list and partition replicas are
> randomly
> >>>>>>>>> distributed to the list.
> >>>>>>>>>
> >>>>>>>>> Thanks,
> >>>>>>>>>
> >>>>>>>>> Jun
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On Wed, Jun 19, 2013 at 10:25 PM, Jason Rosenberg <
> >> j...@squareup.com
> >>>>>>
> >>>>>>>>> wrote:
> >>>>>>>>>
> >>>>>>>>>> In the 0.8 config, log.dir is now log.dirs.  It looks like the
> >>>>>>> singular
> >>>>>>>>>> log.dir is still supported, but under the covers the property is
> >>>>>>>>> log.dirs.
> >>>>>>>>>>
> >>>>>>>>>> I'm curious, does this take a comma separated list of
> directories?
> >>>>>>> The
> >>>>>>>>> new
> >>>>>>>>>> config page just says:
> >>>>>>>>>> "The directories in which the log data is kept"
> >>>>>>>>>>
> >>>>>>>>>> Also, how does kafka handle multiple directories?  Does it treat
> >>>>> each
> >>>>>>>>>> directory as a separate replica partition, or what?
> >>>>>>>>>>
> >>>>>>>>>> Jason
> >>
>

Reply via email to