I'm not familiar with PyCharm. But if you can run "pyspark" from the command line and not hit this, then this might be an issue with PyCharm or your environment - e.g. having an old version of the pyspark code around, or maybe PyCharm itself might need to be updated.
On Thu, Jun 14, 2018 at 10:01 PM, Aakash Basu <aakash.spark....@gmail.com> wrote: > Hi, > > Downloaded the latest Spark version because the of the fix for "ERROR > AsyncEventQueue:70 - Dropping event from queue appStatus." > > After setting environment variables and running the same code in PyCharm, > I'm getting this error, which I can't find a solution of. > > Exception in thread "main" java.util.NoSuchElementException: key not found: > _PYSPARK_DRIVER_CONN_INFO_PATH > at scala.collection.MapLike$class.default(MapLike.scala:228) > at scala.collection.AbstractMap.default(Map.scala:59) > at scala.collection.MapLike$class.apply(MapLike.scala:141) > at scala.collection.AbstractMap.apply(Map.scala:59) > at > org.apache.spark.api.python.PythonGatewayServer$.main(PythonGatewayServer.scala:64) > at > org.apache.spark.api.python.PythonGatewayServer.main(PythonGatewayServer.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > Any help? > > Thanks, > Aakash. -- Marcelo --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org