Hi,
i am trying to read files which are ISO-8859-6 encoded via spark streaming,
but the default encoding for
" ssc.textFileStream " is UTF-8 , so i don't get the data properly , so is
there a way change the default encoding for textFileStream , or a way to
read the file's bytes then i can handl
i am doing word count example on flume stream and trying to save output as
text files in HDFS , but in the save directory i got multiple sub
directories each having files with small size , i wonder if there is a way
to append in a large file instead of saving in multiple files , as i intend
to save
i was having a similar issue but it was in spark and flume integration i was
getting failed to bind error , but got it fixed by shutting down firewall
for both machines (make sure : service iptables status => firewall stopped)
--
View this message in context:
http://apache-spark-user-list.1001
have you fixed this issue ?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Streaming-on-Yarn-Input-from-Flume-tp11755p22055.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
please if you have found a solution for this , could you please post it ?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Store-DStreams-into-Hive-using-Hive-Streaming-tp18307p21877.html
Sent from the Apache Spark User List mailing list archive at Nabble.com
I am trying to store my word count output into hive data warehouse my
pipeline is:
Flume streaming => spark do word count => store result in hive table for
visualization later
my code is :
*import org.apache.spark.SparkContext
import org.apache.spark.SparkContext._
import org.apache.spark.SparkC