Re: installing packages with pyspark

2016-03-19 Thread Felix Cheung
Subject: Re: installing packages with pyspark To: Ajinkya Kale Cc: Hi, regarding 1, packages are resolved locally. That means that when you specify a package, spark-submit will resolve the dependencies and download any jars on the local machine, before shipping* them

Re: installing packages with pyspark

2016-03-19 Thread Franc Carter
k packages, you could also use > --packages in the command line of spark-submit or pyspark. See > http://spark.apache.org/docs/latest/submitting-applications.html > > _ > From: Jakob Odersky > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing

Re: installing packages with pyspark

2016-03-19 Thread Felix Cheung
that are Spark packages, you could also use > --packages in the command line of spark-submit or pyspark. See > http://spark.apache.org/docs/latest/submitting-applications.html > > _ > From: Jakob Odersky > Sent: Thursday, March 17, 2016 6:40 PM > Subjec

Re: installing packages with pyspark

2016-03-19 Thread Ajinkya Kale
___ > From: Jakob Odersky > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing packages with pyspark > To: Ajinkya Kale > Cc: > > > Hi, > regarding 1, packages are resolved locally. That means that when you > specify a package, spark-submit will resolv

Re: installing packages with pyspark

2016-03-19 Thread Jakob Odersky
Hi, regarding 1, packages are resolved locally. That means that when you specify a package, spark-submit will resolve the dependencies and download any jars on the local machine, before shipping* them to the cluster. So, without a priori knowledge of dataproc clusters, it should be no different to

Re: installing packages with pyspark

2016-03-19 Thread Franc Carter
/submitting-applications.html > > _ > From: Jakob Odersky > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing packages with pyspark > To: Ajinkya Kale > Cc: > > > > Hi, > regarding 1, packages are resolved locally. That means

Re: installing packages with pyspark

2016-03-19 Thread Jakob Odersky
/spark.apache.org/docs/latest/submitting-applications.html >> >> _ >> From: Jakob Odersky >> Sent: Thursday, March 17, 2016 6:40 PM >> Subject: Re: installing packages with pyspark >> To: Ajinkya Kale >> Cc: >> >> >&g

installing packages with pyspark

2016-03-19 Thread Ajinkya Kale
Hi all, I had couple of questions. 1. Is there documentation on how to add the graphframes or any other package for that matter on the google dataproc managed spark clusters ? 2. Is there a way to add a package to an existing pyspark context through a jupyter notebook ? --aj