Unsubscribe
Thanks & Best Regards,
Engr. Palash Gupta
Consultant, OSS/CEM/Big Data
Skype: palash2494
https://www.linkedin.com/in/enggpalashgupta
Thanks & Best Regards,
Engr. Palash Gupta
Consultant, OSS/CEM/Big Data
Skype: palash2494
https://www.linkedin.com/in/enggpalashgupta
ask.run(Task.scala:85)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
... 1 more
17/02/08 07:03:51 INFO spark.SparkContext: Invoking stop() from shutdown hook
Thanks & Best Regards,
Palash Gupta
Hi,
You need to add mode overwrite option to avoid this error.
//P.Gupta
Sent from Yahoo Mail on Android
On Fri, 20 Jan, 2017 at 2:15 am, VND Tremblay, Paul
wrote:
I have come across a problem when writing CSV files to S3 in Spark 2.02. The
problem does not exist in Spark 1.6.
19:0
32
Later, I read that the term "cores" doesn't mean physical CPU cores but rather
#tasks that an executor can execute.
Anyway, I don't have a clear idea how to set the number of executors per
physical node. I see there's an option in the Yarn mode, but it's not avai
Hi,
Can you please share how you are assigning cpu core & tell us spark version and
language you are using?
//Palash
Sent from Yahoo Mail on Android
On Wed, 18 Jan, 2017 at 10:16 pm, Saliya Ekanayake wrote:
Thank you, for the quick response. No, this is not Spark SQL. I am running the
bu
ase create a collection with a
record. Otherwise mongo may not keep that db if online session die
//Palash
Sent from Yahoo Mail on Android
On Tue, 17 Jan, 2017 at 12:44 pm, Palash Gupta
wrote: Hi Marco,
What is the user and password you are using for mongodb connection? Did you
enable authori
Hi Marco,
What is the user and password you are using for mongodb connection? Did you
enable authorization?
Better to include user & pass in mongo url.
I remember I tested with python successfully.
Best Regards,Palash
Sent from Yahoo Mail on Android
On Tue, 17 Jan, 2017 at 5:37 am, Marco Mis
Hello Mr. Burton,
Can you share example code how did you implement for other user to see?
"So I think what we did is did a repartition too large and now we ran out of
memory in spark shell. "
Thanks!
P.Gupta
Sent from Yahoo Mail on Android
On Tue, 10 Jan, 2017 at 8:20 am, Kevin Burton wrote:
quot;8g").config("spark.executor.memory",
"4g").appName(APP_NAME).getOrCreate()
Thanks & Best Regards,
Engr. Palash GuptaWhatsApp/Viber: +8801817181502Skype: palash2494
Thanks & Best Regards,Palash Gupta
Hi Macro,
Yes it was in the same host when problem was found.
Even when I tried to start with different host, the problem is still there.
Any hints or suggestion will be appreciated.
Thanks & Best Regards,
Palash Gupta
From: Marco Mistroni
To: Palash Gupta
Cc: ayan guha ; User
File "/usr/local/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63,
in deco
File "/usr/local/spark/python/lib/py4j-0.10.1-src.zip/py4j/protocol.py", line
312, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling o58.load.
Thanks
k application at a time and try to
reproduce this error for both file system and HDFS loading case
When I'm done I will share details with you.
If you have any suggestion more for debug point of view, you can add here for me
Thanks & Best Regards,
Palash Gupta
From: Marco Mis
Hi,
If you want to load from csv, you can use below procedure. Of course you need
to define spark context first. (Given example to load all csv under a folder,
you can use specific name for single file)
// these lines are equivalent in Spark 2.0
spark.read.format("csv").option("header", "true").l
nk what Macro meant by "reduce functionality" is to reduce scope
of your application's functionality so that you can isolate the issue in
certain part(s) of the app...I do not think he meant "reduce" operation :)
On Fri, Dec 30, 2016 at 9:26 PM, Palash Gupta wrote:
Hi
if you
provide some example of reduce functionality. Cause I'm using Spark data frame,
join data frames, use SQL statement to manipulate KPI(s). Here How could I
apply reduce functionality?
Thanks & Best Regards,
Palash Gupta
From: Marco Mistroni
To: "spline_pal...
loads/*.csv")
spark.read.option("header", "true").csv("../Downloads/*.csv") Thanks & Best
Regards,
Palash Gupta
From: Nicholas Hakobian
To: "spline_pal...@yahoo.com"
Cc: Marco Mistroni ; User
Sent: Thursday, December 29, 2016 10:39 P
you can read it
successfully then your hadoop file is the issue and you can start debugging
from there.Hth
On 29 Dec 2016 6:26 am, "Palash Gupta" wrote:
Hi Apache Spark User team,
Greetings!
I started developing an application using Apache Hadoop and Spark using python.
My pyspa
Hi Apache Spark User team,
Greetings!
I started developing an application using Apache Hadoop and Spark using python.
My pyspark application randomly terminated saying "Failed to get broadcast_1*"
and I have been searching for suggestion and support in Stakeoverflow at Failed
to get broadcast_
19 matches
Mail list logo