Testing out 0.6.1 with Spark 2.0 and discovered the way we load dependencies doesn't seem to be working with the new update.
We pass new dependencies in via a SPARK_SUBMIT_OPTIONS environment variable pass the following flags: --packages com.databricks:spark-avro_2.11:3.0.0 Now when I try to import it with: import com.databricks.spark.avro._ I get: <console>:25: error: object databricks is not a member of package com import com.databricks.spark.avro._ I checked the logs are there is no error retrieving the package. So it seems to be something with the classpath. This works in 0.6.0. Any idea if something changed or if we're doing something wrong? I tried this with a few internal packages as well and it doesn't work with those either. Thanks, Mike