ectory on my laptop itself.
Am I crazy? Perhaps this isn't a supported way to use Spark? Any help
or insights are much appreciated!
-Ryan Victory
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: +
re are 6 columns and 4 records in total. These are the sample records.
Should I load it as RDD and then may be using a regex should eliminate
the new lines? Or how it should be? with ". /n" ?
Any suggestions?
Thanks,
Sid
--
Apostolos N. Papadopoulos, Associate Professor
Depar
"true").option("quote",
'"').option(
"delimiter", ",").csv("path")
What else I can do?
Thanks,
get Spark to work on my laptop.
Michael Martin
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email:papad...@csd.auth.gr
twitter: @papadopoulos_ap
web:http://datalab.csd.auth.gr/~apostol
culate the exact partitions needed to load a
specific file?
Thanks,
Sid
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email: papad...@csd.auth.gr
twitter: @papadopoulos_ap
web
/github.com/JerryLead/SparkInternals/blob/master/markdown/english/4-shuffleDetails.md>.
How then is this an improvement on map-reduce?
Image from https://youtu.be/7ooZ4S7Ay6Y
thanks!
--
Apostolos N. Papadopoulos, Associate Professor
Depar
ctions.col(columnNames[i]).isNotNull()).select(columnNames[i]).distinct().collectAsList();
for (int j=0;jcolumnList.add(columnValues.get(j).apply(0).toString());
finalList.add(columnList);|
How to improve this?
Also, can I get the results in JSON format?
--
Apostolos N. Papadopoulos, Associate
560.n3.nabble.com/
-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++003031231099
t() with two columns ("Column1", "Column2"), and i want to
append rows dynamically in a for loop.
Is there any way to achieve this?
Thank you in advance.
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloni
0.n3.nabble.com/
-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email: papad...@csd.
the error. The Error is
in the attachement Pyspark_Error.txt.
Could you please explain me what is this error and how to overpass it?
-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org
--
Apostolos N. Papadopoulos
Thanks
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email: papad...@csd.auth.gr
twitter: @papadopoulos_ap
web: http://dat
driver program. If I can write data to hdfs at executor, then the driver memory
for my spark job can be reduced.
Otherwise does Spark support streaming read from database (i.e. spark streaming
+ spark sql)?
Thanks for your reply.
‐‐‐ Original Message ‐‐‐
On 7 September 2018 4:15 PM
On 14/09/2018 11:21 πμ, Aakash Basu wrote:
Hi,
What is the Spark cluster equivalent of standalone's local[N]. I mean,
the value we set as a parameter of local as N, which parameter takes
it in the cluster mode?
Thanks,
Aakash.
--
Apostolos N. Papadopoulos, Associate Professor
Departme
x27;)
,('spark.network.timeout', '800')
,('spark.scheduler.mode', 'FAIR')
,('spark.shuffle.service.enabled', 'true')
,('spark.dynamicAllocation.enabled', 'true')
])
py_files =
['hdfs://emr-header-1.
or.
How can i apply Kmeans to every partition?
Thank you in advance,
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email: papad...@csd.auth.gr
twitter: @papadopoulos_ap
web: http:/
at
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Thanks,
Asmath
--
Apostolos N. Papadopoulos, Associate Professor
Dep
n.
Any clue is highly appreciated! Thanks.
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle University of Thessaloniki
Thessaloniki, GREECE
tel: ++0030312310991918
email: papad...@csd.auth.gr
twitter: @papadopoulos_ap
web:
nce of
java.lang.invoke.SerializedLambda to field
org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in
instance of org.apache.spark.rdd.MapPartitionsRDD
I was wondering if anyone has seen this before.
Thanks
Nikhil
--
Apostolos N. Papadopoulos, Associate Professor
Department of Informatics
Aristotle Univers
19 matches
Mail list logo