If you want to submit applications to a remote cluster where your port 7077
is opened publically, then you would need to set the *spark.driver.host *(with
the public ip of your laptop) and *spark.driver.port* (optional, if there's
no firewall between your laptop and the remote cluster). Keeping your 7077
open for public is a bad idea, you can read more here
https://www.sigmoid.com/securing-apache-spark-cluster/

Thanks
Best Regards

On Mon, Jun 1, 2015 at 11:48 PM, AlexG <swift...@gmail.com> wrote:

> I've followed the instructions for setting up a standalone spark cluster
> (on
> EC2):
> - install spark on all the machines
> - enabled passwordless ssh
> - setup the conf/slaves file
> - start the master and slaves with the provided scripts
>
> The status on the 8080 port of the master tells me that the master and
> executors are all running. I can successfully use pyspark from the master.
>
> However, if I try to call pyspark remotely from my laptop, with
> MASTER=spark://<ip>:7077 pyspark,
> I get these errors:
> 15/06/01 10:02:14 INFO AppClient$ClientActor: Connecting to master
> akka.tcp://sparkMaster@<IP>:7077/user/Master...
> 15/06/01 10:02:34 INFO AppClient$ClientActor: Connecting to master
> akka.tcp://sparkMaster@<IP>:7077/user/Master...
> 15/06/01 10:02:54 INFO AppClient$ClientActor: Connecting to master
> akka.tcp://sparkMaster@<IP>:7077/user/Master...
> 15/06/01 10:03:14 ERROR SparkDeploySchedulerBackend: Application has been
> killed. Reason: All masters are unresponsive! Giving up.
> 15/06/01 10:03:14 ERROR TaskSchedulerImpl: Exiting due to error from
> cluster
> scheduler: All masters are unresponsive! Giving up.
>
> Any idea what's going on here? I set port 7077 to be publicly accessible...
>
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/using-pyspark-with-standalone-cluster-tp23099.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to