[ 
https://issues.apache.org/jira/browse/SPARK-18150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Owen resolved SPARK-18150.
-------------------------------
    Resolution: Invalid

Please start on the mailing list with a more detailed question, and after 
reviewing the contributing guide.

>  Spark 2.* failes to create partitions for avro files
> -----------------------------------------------------
>
>                 Key: SPARK-18150
>                 URL: https://issues.apache.org/jira/browse/SPARK-18150
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL, Streaming
>            Reporter: Sunil Kumar
>            Priority: Blocker
>
> I am using Apache Spark 2.0.1 for processing the Grid HDFS Avro file, however 
> I don't see spark distributing the job into different tasks instead it uses 
> single task and all the operations (read, load, filter, show ) are done in a 
> sequence using same task.
> This means I am not able to leverage distributed parallel processing.
> I tried the same operation on JSON file on HDFS, it works good, means the job 
> gets distributed into multiple tasks and partition. I see parallelism.
> I then tested the same on Spark 1.6, there it does the partitioning. Looks 
> like there is a bug in Spark 2.* version. If not can some one help me know 
> how to achieve the same on Avro file, do I need to do something special for 
> Avro files ?
> Note:
> I explored spark setting: "spark.default.parallelism",  
> "spark.sql.files.maxPartitionBytes", "--num-executors" and 
> "spark.sql.shuffle.partitions". These were not of much help. 
> "spark.default.parallelism", ensured to have multiple tasks however a single 
> task ended up performing all the operation.
> I am using com.databricks.spark.avro (3.0.1) for Spark 2.0.1.
> Thanks,
> Sunil



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to