*Regards,*
*Mahebub Sayyed*
eference* program to stream data from kafka
>> topics and output as parquet file on HDFS.
>>
>> Thanks in Advance.
>>
>> Regards,
>>
>> Rafeeq S
>> *(“What you do is what matters, not what you think or say or plan.” )*
>>
>>
>
--
*Regards,*
*Mahebub Sayyed*
provided
org.apache.spark
spark-streaming-kafka_2.10
1.0.1
Please help me.
--
*Regards,*
*Mahebub Sayyed*
> >
> > I am new user of spark, I am using cloudera-quickstart-vm-5.0.0-0-vmware
> for
> > execute sample examples of Spark.
> > I am very sorry for silly and basic question.
> > I am not able to deploy and execute sample examples of spark.
> >
> > please suggest me how to start with spark.
> >
> > Please help me
> > Thanks in advance.
> >
> > Regards,
> > Sam
> >
> >
>
--
*Regards,*
*Mahebub Sayyed*
ataset (as RDDs) to two
> different parquet files, you just have to call saveAsParquetFile twice (on
> same or transformed versions of the RDD), as shown in the guide.
>
> Hope this helps!
>
> TD
>
>
> On Thu, Jul 17, 2014 at 2:19 AM, Mahebub Sayyed
> wrote:
>
>>
Hi All,
Currently we are reading (multiple) topics from Apache kafka and storing
that in HBase (multiple tables) using twitter storm (1 tuple stores in 4
different tables).
but we are facing some performance issue with HBase.
so we are replacing* HBase* with *Parquet* file and *storm* with *Apache
me.
Thanks in advance.
--
*Regards,*
*Mahebub Sayyed*