I believe either should work. The broker has a record of what it should have in zk and will recreate any missing logs. Try it to make sure though.
Sent from my iPhone On Aug 15, 2013, at 12:52 AM, Jason Rosenberg <j...@squareup.com> wrote: > Ok, that makes sense that the broker will shut itself down. > > If we bring it back up, can this be with an altered set of log.dirs? Will > the destroyed partitions get rebuilt on a new log.dir? Or do we have to > bring it back up with a new or repaired disk, matching the old log.dir, in > order for those replicas to be rebuilt? > > Jason > > > On Wed, Aug 14, 2013 at 4:16 PM, Jay Kreps <jay.kr...@gmail.com> wrote: > >> If you get a disk error that results in an IOException the broker will >> shut itself down. You would then have the option of replacing the disk or >> deleting that data directory from the list. When the broker is brought back >> up the intact partitions will quickly catch up and be online; the destroyed >> partitions will have to fully rebuild off the other replicas and will take >> a little longer but will automatically come back online once they have >> restored off the replicas. >> >> -jay >> >> Sent from my iPhone >> >> On Aug 14, 2013, at 1:49 PM, Jason Rosenberg <j...@squareup.com> wrote: >> >>> I'm getting ready to try out this configuration (use multiple disks, no >>> RAID, per broker). One concern is the procedure for recovering if there >> is >>> a disk failure. >>> >>> If a disk fails, will the broker go offline, or will it continue serving >>> partitions on its remaining good disks? And if so, is there a procedure >>> for moving the partitions that were on the failed disk, but not >> necessarily >>> all the others on that broker? >>> >>> Jason >>> >>> >>> On Thu, Jun 20, 2013 at 3:15 PM, Jason Rosenberg <j...@squareup.com> >> wrote: >>> >>>> yeah, that would work! >>>> >>>> >>>> On Thu, Jun 20, 2013 at 1:20 PM, Jay Kreps <jay.kr...@gmail.com> wrote: >>>> >>>>> Yeah we didn't go as far as adding weighting or anything like that--I >>>>> think we'd be open to a patch that did that as long as it was >>>>> optional. In the short term you can obviously add multiple directories >>>>> on the same disk to increase its share. >>>>> >>>>> -Jay >>>>> >>>>> On Thu, Jun 20, 2013 at 12:59 PM, Jason Rosenberg <j...@squareup.com> >>>>> wrote: >>>>>> This sounds like a great idea, to just disks as "just a bunch of >> disks" >>>>> or >>>>>> JBOD.....hdfs works well this way. >>>>>> >>>>>> Do all the disks need to be the same size, to use them evenly? Since >> it >>>>>> will allocate partitions randomly? >>>>>> >>>>>> It would be nice if you had 2 disks, with one twice as large as the >>>>> other, >>>>>> if the larger would be twice as likely to receive partitions as the >>>>> smaller >>>>>> one, etc. >>>>>> >>>>>> I suppose this goes into my earlier question to the list, vis-a-vis >>>>>> heterogeneous brokers (e.g. utilize brokers with different sized >>>>> storage, >>>>>> using some sort of weighting scheme, etc.). >>>>>> >>>>>> Jason >>>>>> >>>>>> >>>>>> On Thu, Jun 20, 2013 at 11:07 AM, Jay Kreps <jay.kr...@gmail.com> >>>>> wrote: >>>>>> >>>>>>> The intention is to allow the use of multiple disks without RAID or >>>>>>> logical volume management. We have found that there are a lot of >>>>>>> downsides to RAID--in particular a huge throughput hit. Since we >>>>>>> already have a parallelism model due to partitioning and a fault >>>>>>> tolerance model with replication RAID doesn't actually buy much. With >>>>>>> this feature you can directly mount multiple disks as their own >>>>>>> directory and the server will randomly assign partitions to them. >>>>>>> >>>>>>> Obviously this will only work well if there are enough >> high-throughput >>>>>>> partitions to make load balance evenly (e.g. if you have only one big >>>>>>> partition per server then this isn't going to work). >>>>>>> >>>>>>> -Jay >>>>>>> >>>>>>> On Wed, Jun 19, 2013 at 11:01 PM, Jason Rosenberg <j...@squareup.com> >>>>>>> wrote: >>>>>>>> is it possible for a partition to have multiple replicas on >> different >>>>>>>> directories on the same broker? (hopefully no!) >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Jun 19, 2013 at 10:47 PM, Jun Rao <jun...@gmail.com> wrote: >>>>>>>> >>>>>>>>> It takes a comma separated list and partition replicas are randomly >>>>>>>>> distributed to the list. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Jun >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Jun 19, 2013 at 10:25 PM, Jason Rosenberg < >> j...@squareup.com >>>>>> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> In the 0.8 config, log.dir is now log.dirs. It looks like the >>>>>>> singular >>>>>>>>>> log.dir is still supported, but under the covers the property is >>>>>>>>> log.dirs. >>>>>>>>>> >>>>>>>>>> I'm curious, does this take a comma separated list of directories? >>>>>>> The >>>>>>>>> new >>>>>>>>>> config page just says: >>>>>>>>>> "The directories in which the log data is kept" >>>>>>>>>> >>>>>>>>>> Also, how does kafka handle multiple directories? Does it treat >>>>> each >>>>>>>>>> directory as a separate replica partition, or what? >>>>>>>>>> >>>>>>>>>> Jason >>