RE: Size exceeds Integer.MAX_VALUE on EMR 4.0.0 Spark 1.4.1

2015-11-16 Thread Ewan Leith
big it would be, I assume it’s over 2 GB From: Zhang, Jingyu [mailto:jingyu.zh...@news.com.au] Sent: 16 November 2015 10:17 To: user Subject: Size exceeds Integer.MAX_VALUE on EMR 4.0.0 Spark 1.4.1 I am using spark-csv to save files in s3, it shown Size exceeds. Please let me know how to fix it

Re: Size exceeds Integer.MAX_VALUE on EMR 4.0.0 Spark 1.4.1

2015-11-16 Thread Sabarish Sasidharan
You can try increasing the number of partitions before writing it out. Regards Sab On Mon, Nov 16, 2015 at 3:46 PM, Zhang, Jingyu wrote: > I am using spark-csv to save files in s3, it shown Size exceeds. Please let > me know how to fix it. Thanks. > > df.write() > .format("com.databricks.s

Size exceeds Integer.MAX_VALUE on EMR 4.0.0 Spark 1.4.1

2015-11-16 Thread Zhang, Jingyu
I am using spark-csv to save files in s3, it shown Size exceeds. Please let me know how to fix it. Thanks. df.write() .format("com.databricks.spark.csv") .option("header", "true") .save("s3://newcars.csv"); java.lang.IllegalArgumentException: Size exceeds Integer.MAX_VALUE at