That project is for reading data in from Redshift table exports stored in s3 by
running commands in redshift like this:
unload ('select * from venue')
to 's3://mybucket/tickit/unload/'
http://docs.aws.amazon.com/redshift/latest/dg/t_Unloading_tables.html
The path in the parameters below is the s3 bucket path.
Hope this helps,
Ewan
-----Original Message-----
From: Hafiz Mujadid [mailto:[email protected]]
Sent: 05 June 2015 15:25
To: [email protected]
Subject: redshift spark
Hi All,
I want to read and write data to aws redshift. I found spark-redshift project
at following address.
https://github.com/databricks/spark-redshift
in its documentation there is following code is written.
import com.databricks.spark.redshift.RedshiftInputFormat
val records = sc.newAPIHadoopFile(
path,
classOf[RedshiftInputFormat],
classOf[java.lang.Long],
classOf[Array[String]])
I am unable to understand it's parameters. Can somebody explain how to use
this? what is meant by path in this case?
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/redshift-spark-tp23175.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected] For additional
commands, e-mail: [email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]