Hi Annabel,

I am using Spark in stand-alone mode (deployment using the ec2 scripts
packaged with spark).

Cheers,
Michael

On 08.01.2016 00:43, Annabel Melongo wrote:
> Michael,
>
> I don't know what's your environment but if it's Cloudera, you should
> be able to see the link to your master in the Hue.
>
> Thanks
>
>
> On Thursday, January 7, 2016 5:03 PM, Michael Pisula
> <michael.pis...@tngtech.com> wrote:
>
>
> I had tried several parameters, including --total-executor-cores, no
> effect.
> As for the port, I tried 7077, but if I remember correctly I got some
> kind of error that suggested to try 6066, with which it worked just
> fine (apart from this issue here).
>
> Each worker has two cores. I also tried increasing cores, again no
> effect. I was able to increase the number of cores the job was using
> on one worker, but it would not use any other worker (and it would not
> start if the number of cores the job wanted was higher than the number
> available on one worker).
>
> On 07.01.2016 22:51, Igor Berman wrote:
>> read about *--total-executor-cores*
>> not sure why you specify port 6066 in master...usually it's 7077
>> verify in master ui(usually port 8080) how many cores are
>> there(depends on other configs, but usually workers connect to master
>> with all their cores)
>>
>> On 7 January 2016 at 23:46, Michael Pisula
>> <michael.pis...@tngtech.com <mailto:michael.pis...@tngtech.com>> wrote:
>>
>>     Hi,
>>
>>     I start the cluster using the spark-ec2 scripts, so the cluster
>>     is in stand-alone mode.
>>     Here is how I submit my job:
>>     spark/bin/spark-submit --class demo.spark.StaticDataAnalysis
>>     --master spark://<host>:6066 --deploy-mode cluster
>>     demo/Demo-1.0-SNAPSHOT-all.jar
>>
>>     Cheers,
>>     Michael
>>
>>
>>     On 07.01.2016 22:41, Igor Berman wrote:
>>>     share how you submit your job
>>>     what cluster(yarn, standalone)
>>>
>>>     On 7 January 2016 at 23:24, Michael Pisula
>>>     <michael.pis...@tngtech.com <mailto:michael.pis...@tngtech.com>>
>>>     wrote:
>>>
>>>         Hi there,
>>>
>>>         I ran a simple Batch Application on a Spark Cluster on EC2.
>>>         Despite having 3
>>>         Worker Nodes, I could not get the application processed on
>>>         more than one
>>>         node, regardless if I submitted the Application in Cluster
>>>         or Client mode.
>>>         I also tried manually increasing the number of partitions in
>>>         the code, no
>>>         effect. I also pass the master into the application.
>>>         I verified on the nodes themselves that only one node was
>>>         active while the
>>>         job was running.
>>>         I pass enough data to make the job take 6 minutes to process.
>>>         The job is simple enough, reading data from two S3 files,
>>>         joining records on
>>>         a shared field, filtering out some records and writing the
>>>         result back to
>>>         S3.
>>>
>>>         Tried all kinds of stuff, but could not make it work. I did
>>>         find similar
>>>         questions, but had already tried the solutions that worked
>>>         in those cases.
>>>         Would be really happy about any pointers.
>>>
>>>         Cheers,
>>>         Michael
>>>
>>>
>>>
>>>         --
>>>         View this message in context:
>>>         
>>> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-job-uses-only-one-Worker-tp25909.html
>>>         Sent from the Apache Spark User List mailing list archive at
>>>         Nabble.com.
>>>
>>>         
>>> ---------------------------------------------------------------------
>>>         To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>         <mailto:user-unsubscr...@spark.apache.org>
>>>         For additional commands, e-mail: user-h...@spark.apache.org
>>>         <mailto:user-h...@spark.apache.org>
>>>
>>>
>>
>>     -- 
>>     Michael Pisula * michael.pis...@tngtech.com 
>> <mailto:michael.pis...@tngtech.com> * +49-174-3180084
>>     TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring
>>     Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke
>>     Sitz: Unterföhring * Amtsgericht München * HRB 135082
>>
>>
>
> -- 
> Michael Pisula * michael.pis...@tngtech.com 
> <mailto:michael.pis...@tngtech.com> * +49-174-3180084
> TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring
> Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke
> Sitz: Unterföhring * Amtsgericht München * HRB 135082
>
>

-- 
Michael Pisula * michael.pis...@tngtech.com * +49-174-3180084
TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring
Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke
Sitz: Unterföhring * Amtsgericht München * HRB 135082

Reply via email to