Re: Spark task hangs infinitely when accessing S3 from AWS

2016-01-26 Thread aecc
Sorry, I have not been able to solve the issue. I used speculation mode as workaround to this. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-task-hangs-infinitely-when-accessing-S3-from-AWS-tp25289p26068.html Sent from the Apache Spark User List mail

Spark task hangs infinitely when accessing S3 from AWS

2015-11-05 Thread aecc
Hi guys, when reading data from S3 from AWS using Spark 1.5.1 one of the tasks hangs when reading data in a way that cannot be reproduced. Some times it hangs, some times it doesn't. This is the thread dump from the hung task: "Executor task launch worker-3" daemon prio=10 tid=0x7f419c023000

Re: Spark task hangs infinitely when accessing S3 from AWS

2015-11-09 Thread aecc
Any help on this? this is really blocking me and I don't find any feasible solution yet. Thanks. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-task-hangs-infinitely-when-accessing-S3-from-AWS-tp25289p25327.html Sent from the Apache Spark User List m

Re: Spark task hangs infinitely when accessing S3 from AWS

2015-11-12 Thread aecc
Any hints? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-task-hangs-infinitely-when-accessing-S3-from-AWS-tp25289p25365.html Sent from the Apache Spark User List mailing list archive at Nabble.com. ---

Re: Spark task hangs infinitely when accessing S3 from AWS

2015-11-12 Thread aecc
Some other stats: The number of files I have in the folder is 48. The number of partitions used when reading data is 7315. The maximum size of a file to read is 14G The size of the folder is around: 270G -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spar

zookeeper mesos logging in spark

2016-08-26 Thread aecc
Hi, Everytime I run my spark application using mesos, I get logs in my console in the form: 2016-08-26 15:25:30,949:960521(0x7f6bccff9700):ZOO_INFO@log_env 2016-08-26 15:25:30,949:960521(0x7f6bccff9700):ZOO_INFO@log_env 2016-08-26 15:25:30,949:960521(0x7f6bccff9700):ZOO_INFO@log_env 2016-08-26 15

Using Spark Context as an attribute of a class cannot be used

2014-11-24 Thread aecc
Hello guys, I'm using Spark 1.0.0 and Kryo serialization In the Spark Shell, when I create a class that contains as an attribute the SparkContext, in this way: class AAA(val s: SparkContext) { } val aaa = new AAA(sc) and I execute any action using that attribute like: val myNumber = 5 aaa.s.tex

Re: Using Spark Context as an attribute of a class cannot be used

2014-11-24 Thread aecc
Marcelo Vanzin wrote > Do you expect to be able to use the spark context on the remote task? Not At all, what I want to create is a wrapper of the SparkContext, to be used only on the driver node. I would like to have in this "AAA" wrapper several attributes, such as the SparkContext and other con

Re: Using Spark Context as an attribute of a class cannot be used

2014-11-24 Thread aecc
If I actually instead of using myNumber I use the 5 value, the exception is not given. E.g: aaa.s.parallelize(1 to 10).filter(_ == 5).count Works perfectly -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Using-Spark-Context-as-an-attribute-of-a-class-canno

Re: Using Spark Context as an attribute of a class cannot be used

2014-11-24 Thread aecc
n3.nabble.com>: > Hello, > > On Mon, Nov 24, 2014 at 12:07 PM, aecc <[hidden email] > <http://user/SendEmail.jtp?type=node&node=19687&i=0>> wrote: > > This is the stacktrace: > > > > org.apache.spark.SparkException: Job aborted due to stage fail

Re: Using Spark Context as an attribute of a class cannot be used

2014-11-24 Thread aecc
eeds to be serialized, and this: sc.parallelize(1 to 10).filter(_ == myNumber).count does not. 2014-11-24 23:13 GMT+01:00 Marcelo Vanzin [via Apache Spark User List] < ml-node+s1001560n19692...@n3.nabble.com>: > On Mon, Nov 24, 2014 at 1:56 PM, aecc <[hidden email] > <http://user/

Locality level and Kryo

2014-12-16 Thread aecc
Hi guys, It happens to me quite often that when the locality level of a task goes further than LOCAL (NODE, RACK, etc), I get some of the following exceptions: "too many files open", "encountered unregistered class id", "cannot cast X to Y". I do not get any exceptions during shuffling (which mea

Locality Level Kryo

2014-12-16 Thread aecc
Hi guys, I get Kryo exceptions of the type "unregistered class id" and "cannot cast to class" when the locality level of the tasks go beyond LOCAL. However I get no Kryo exceptions during shuffling operations. If the locality level never goes beyond LOCAL everything works fine. Is there a speci

Spark in YARN HDP problem

2014-02-25 Thread aecc
Hi, I'm trying to run the Spark examples in YARN and I get the following error: appDiagnostics: Application application_1390483691679_0124 failed 2 times due to AM Container for appattempt_1390483691679_0124_02 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoo

Re: Spark in YARN HDP problem

2014-02-26 Thread aecc
The error happens after the application is accepted: 14/02/26 12:05:08 INFO YarnClientImpl: Submitted application application_1390483691679_0157 to ResourceManager at X:8050 3483 [main] INFO org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend - Application report from ASM:

Re: Spark in YARN HDP problem

2014-02-26 Thread aecc
Hi Azuryy, This is what I got: LogType: stderr LogLength: 85 Log Contents: Error: Could not find or load main class org.apache.spark.deploy.yarn.WorkerLauncher LogType: stdout LogLength: 0 Log Contents: That made me thought about the consolidated jar file that I was including. It was a simple

Kafka in Yarn

2014-03-13 Thread aecc
Hi, I would like to know how is the correct way to add kafka to my project in StandAlone YARN, given that now it's in a different artifact than the Spark core. I tried adding the dependency to my project but I get a NotClassFoundException to my main class. Also, that makes my Jar file very big, so

Sliding Subwindows

2014-04-01 Thread aecc
Hello, I would like to have a kind of sub windows. The idea is to have 3 windows in the following way: future <> <-> <--> past w1 w2 w3 So I can do some processing with