Hi Colin, I have a requirement to be able to run all the Hadoop functionality that I build from a variety of platforms. This was the original motivation for wanting to use libhdfs. I followed your suggestion and looked into using webhdfs and it's looking promising. Thanks for that. However I also need to be able to launch Map/Reduce jobs from any platform. In particular from Windows. I looked into this by hacking the bin/hadoop script to extract the required class path and various arguments so that I could launch a Map/Reduce job just by invoking java with the correct arguments. However I ran into HADOOP-7682. I can see that there is a workaround here https://github.com/congainc/patch-hadoop_7682-1.0.x-win but it suggests that this is not really appropriate for deployment. I suspect that I can get it to work reliably by using cygwin and making loads of modifications but that all seems rather a large effort, error-prone and difficult to maintain.
Given that I plan to have a relatively small repertoire of Map/Reduce jobs that I need to launch, I'm tempted to have all the jars pre-packed on the Name-Node and have the ability to run them there. Then have a daemon running so that I can use any appropriate ad-hoc RPC mechanism from Windows to launch them. Am I missing something? Is there a way to launch Map/Reduce jobs in a platform neutral way, which runs out of the box, on Windows? Again, any suggestions welcome. Peter Marron > -----Original Message----- > From: Peter Marron > [mailto:Peter.Marron@trilliumsoftware. > com] > Sent: 26 October 2012 00:53 > To: common-dev@hadoop.apache.org > Subject: RE: libhdfs on windows > > Hi Colin, > > OK, I didn't know there was a hdfs-dev. > I'm happy to ask there. > (However there's a lot of mail on > dev@hadoop and user@hadoop as well > as user@hive and it's a bit of a > commitment to track them all.) As for > webhdfs, I did think about that, and in > some ways it's a beautiful solution as it > gives me a platform- and language- > neutral access mechanism. I was just a > little worried about the HTTP overhead if > I am reading a single record at a time. > Also I will need some way to launch my > Map/Reduce jobs as well. So I'll probably > end up using the C++/JNI/Java route to > do that anyway. Unless there's a better > way? > Is there a web Map/Reduce interface? > > Many thanks, > > Z > > > -----Original Message----- > > From: rarecac...@gmail.com > > [mailto:rarecac...@gmail.com] On > > Behalf Of Colin McCabe > > Sent: 25 October 2012 18:24 > > To: common-dev@hadoop.apache.org > > Subject: Re: libhdfs on windows > > > > Hi Peter, > > > > This might be a good question for hdfs- > dev? > > > > As Harsh pointed out below, HDFS-573 > > was never committed. I don't even see > a patch attached, although > > there is some discussion. > > > > In the mean time, might I suggest using > the webhdfs interface on > > Windows? > > webhdfs was intended as a stable REST > > interface that can be accessed from > any platform. > > > > cheers, > > Colin > > > > > > On Thu, Oct 25, 2012 at 7:19 AM, Peter > Marron > > <peter.mar...@trilliumsoftware.com> > > wrote: > > > Hi, > > > > > > I've been looking at using libhdfs and > I > > would like to use it on windows. > > > I have found HDFS-573 and the > > information on this page: > > > > > > http://issues.apache.org/jira/browse/HD > > FS-573?page=com.atlassian.jira. > > > plugin.system.issuetabpanels:all- > > tabpanel > > > which suggests that quite a lot of > work > > was done on this way back in 2009. > > > So is there some source from this > > effort retained somewhere? If so, > > where? > > > Or do I have to start from scratch? > > Apologies if this has already been asked > recently. > > > > > > Any help appreciated. > > > > > > Peter Marron > >