Hi ,
Released latest version of Receiver based Kafka Consumer for Spark
Streaming.
Receiver is compatible with Kafka versions 0.8.x, 0.9.x and 0.10.x and All
Spark Versions
Available at Spark Packages :
https://spark-packages.org/package/dibbhatt/kafka-spark-consumer
Also at github : https://g
Hi,
Thanks to Jerry for mentioning the Kafka Spout for Trident. The Storm
Trident has done the exact-once guarantee by processing the tuple in a
batch and assigning same transaction-id for a given batch . The replay for
a given batch with a transaction-id will have exact same set of tuples and
re
Dear All ,
I have been playing with Spark Streaming on Tachyon as the OFF_HEAP block
store . Primary reason for evaluating Tachyon is to find if Tachyon can
solve the Spark BlockNotFoundException .
In traditional MEMORY_ONLY StorageLevel, when blocks are evicted , jobs
failed due to block not fo
-Dtachyon.worker.hierarchystore.level1.dirs.path=/mnt/tachyon
-Dtachyon.worker.hierarchystore.level1.dirs.quota=50GB
-Dtachyon.worker.allocate.strategy=MAX_FREE
-Dtachyon.worker.evict.strategy=LRU
Regards,
Dibyendu
On Thu, May 7, 2015 at 1:46 PM, Dibyendu Bhattacharya <
dibyendu.bhatt
m interface, is returning zero.
>
> On Mon, May 11, 2015 at 4:38 AM, Dibyendu Bhattacharya <
> dibyendu.bhattach...@gmail.com> wrote:
>
>> Just to follow up this thread further .
>>
>> I was doing some fault tolerant testing of Spark Streaming with Tachyon
>>
ne go about configuring spark streaming to use tachyon as its
> place for storing checkpoints? Also, can one do this with tachyon running
> on a completely different node than where spark processes are running?
>
> Thanks
> Nikunj
>
>
> On Thu, May 21, 2015 at 8:35 PM, Dib
Hi,
I have raised a JIRA ( https://issues.apache.org/jira/browse/SPARK-11045)
to track the discussion but also mailing dev group for your opinion. There
are some discussions already happened in Jira and love to hear what others
think. You can directly comment against the Jira if you wish.
This ka
Hi,
I have implemented a Low Level Kafka Consumer for Spark Streaming using
Kafka Simple Consumer API. This API will give better control over the Kafka
offset management and recovery from failures. As the present Spark
KafkaUtils uses HighLevel Kafka Consumer API, I wanted to have a better
control
l.com
>> +1 (206) 849-4108
>>
>>
>> On Sun, Aug 3, 2014 at 8:59 PM, Patrick Wendell
>> wrote:
>>
>>> I'll let TD chime on on this one, but I'm guessing this would be a
>>> welcome addition. It's great to see community effort on adding new
&
ers. I’m not sure what is your thought?
>
> Thanks
> Jerry
>
> From: Dibyendu Bhattacharya [mailto:dibyendu.bhattach...@gmail.com]
> Sent: Tuesday, August 05, 2014 5:15 PM
> To: Jonathan Hodges; dev@spark.apache.org
> Cc: user
> Subject: Re: Low Level Kafka Consumer
understand the details, but I want to do it really soon. In particular, I
> want to understand the improvements, over the existing Kafka receiver.
>
> And its fantastic to see such contributions from the community. :)
>
> TD
>
>
> On Tue, Aug 5, 2014 at 8:38 AM, Dibyendu Bhattach
oolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:744)
>
>
>
>
> --
> Nan Zhu
>
> On Thursday, September 11, 2014 at 10:42 AM, Nan Zhu wrote:
>
&g
So you have a single Kafka topic which has very high retention period (
that decides the storage capacity of a given Kafka topic) and you want to
process all historical data first using Camus and then start the streaming
process ?
The challenge is, Camus and Spark are two different consumer for Ka
13 matches
Mail list logo