Here's
<https://github.com/twitter/hadoop-lzo/blob/master/src/main/java/com/hadoop/mapreduce/LzoTextInputFormat.java>
the class. You can read more here
https://github.com/twitter/hadoop-lzo#maven-repository

Thanks
Best Regards

On Thu, May 14, 2015 at 1:22 PM, lisendong <lisend...@163.com> wrote:

> LzoTextInputFormat where is this class?
> what is the maven dependency?
>
>
> 在 2015年5月14日,下午3:40,Akhil Das <ak...@sigmoidanalytics.com> 写道:
>
> That's because you are using TextInputFormat i think, try
> with LzoTextInputFormat like:
>
> val list_join_action_stream = ssc.fileStream[LongWritable, Text,
> com.hadoop.mapreduce.LzoTextInputFormat](gc.input_dir, (t: Path) => true,
> false).map(_._2.toString)
>
> Thanks
> Best Regards
>
> On Thu, May 14, 2015 at 1:04 PM, lisendong <lisend...@163.com> wrote:
>
>> I have action on DStream.
>> because when I put a text file into the hdfs, it runs normally, but if I
>> put a lz4 file, it does nothing.
>>
>> 在 2015年5月14日,下午3:32,Akhil Das <ak...@sigmoidanalytics.com> 写道:
>>
>> What do you mean by not detected? may be you forgot to trigger some
>> action on the stream to get it executed. Like:
>>
>> val list_join_action_stream = ssc.fileStream[LongWritable, Text,
>> TextInputFormat](gc.input_dir, (t: Path) => true,
>> false).map(_._2.toString)
>>
>> *list_join_action_stream.count().print()*
>>
>>
>>
>>
>> Thanks
>> Best Regards
>>
>> On Wed, May 13, 2015 at 7:18 PM, hotdog <lisend...@163.com> wrote:
>>
>>> in spark streaming, I want to use fileStream to monitor a directory. But
>>> the
>>> files in that directory are compressed using lz4. So the new lz4 files
>>> are
>>> not detected by the following code. How to detect these new files?
>>>
>>>     val list_join_action_stream = ssc.fileStream[LongWritable, Text,
>>> TextInputFormat](gc.input_dir, (t: Path) => true,
>>> false).map(_._2.toString)
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://apache-spark-user-list.1001560.n3.nabble.com/how-to-read-lz4-compressed-data-using-fileStream-of-spark-streaming-tp22868.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com
>>> <http://nabble.com/>.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>
>>>
>>
>>
>
>

Reply via email to