https://issues.apache.org/jira/browse/SPARK-8597
A JIRA ticket discussing the same problem (with more insights than here)!
--
View this message in context:
http://apache-spark-developers-list.1001551.n3.nabble.com/DataFrame-partitionBy-issues-tp12838p12974.html
Sent from the Apache Spark
.partitionBy("A", "B").mode("overwrite").parquet("table.parquet”)
>
> How I ran the Spark shell:
> bin/spark-shell --driver-memory 16g --master local[8] --packages
> com.databricks:spark-csv_2.10:1.0.3
>
> Attached you'll find table.csv which I used
rquet("table.parquet”)
How I ran the Spark shell:
bin/spark-shell --driver-memory 16g --master local[8] --packages
com.databricks:spark-csv_2.10:1.0.3
Attached you'll find table.csv which I used. table.csv
<http://apache-spark-developers-list.1001551.n3.nabble.com/file/n12838/table.csv>
Thank yo