Hi,

Since you didn't get an answer... yes you can.

I'm working from memory so I may be a bit fuzzy on the details...

Your external app has to be 'cloud aware'. Essentially create a config file for 
your application that you can read in which lets your app know where the JT and 
NN are.

Then you can use the Tool (?interface/class) implement/extend with your job 
class and tell it to run...

Sorry for being terse... I'm under the gun on getting something out the door.

HTH

-Mike


-----Original Message-----
From: Sebastian Ruff (open4business GmbH) [mailto:r...@open4business.de] 
Sent: Wednesday, July 28, 2010 7:27 AM
To: common-dev@hadoop.apache.org
Subject: Starting a job on a hadoop cluster remotly

Hey,

 

is it possible to start a job on a hadoop cluster from remote. For
example we have a web application

which runs on an apache tomcat server. And would like to start a
mapreduce job on our cluster, from

within the webapp.

 

Is this possible? And if yes, what are the steps to get there? Do I just
have to put my namenode and datanode

in a core-site.xml in the webapp and call the api?

 

Thanks a lot,

 

Sebastian



The information contained in this communication may be CONFIDENTIAL and is 
intended only for the use of the recipient(s) named above.  If you are not the 
intended recipient, you are hereby notified that any dissemination, 
distribution, or copying of this communication, or any of its contents, is 
strictly prohibited.  If you have received this communication in error, please 
notify the sender and delete/destroy the original message and any copy of it 
from your computer or paper files.

Reply via email to