Hi,
Is that command available for all nodes? Did you try as below? ;)
Process proc = rt.exec("/bin/hostname");
..
output.collect(hostname, disk usage);
On Tue, Apr 28, 2009 at 6:13 PM, Razen Al Harbi <[email protected]> wrote:
> Hi all,
>
> I am writing an application in which I create a forked process to execute a
> specific Map/Reduce job. The problem is that when I try to read the output
> stream of the forked process I get nothing and when I execute the same job
> manually it starts printing the output I am expecting. For clarification I
> will go through the simple code snippet:
>
>
> Process p = rt.exec("hadoop jar GraphClean args");
> BufferedReader reader = new BufferedReader(new
> InputStreamReader(p.getInputStream()));
> String line = null;
> check = true;
> while(check){
> line = reader.readLine();
> if(line != null){// I know this will not finish it's only for testing.
> System.out.println(line);
> }
> }
>
> If I run this code nothing shows up. But if execute the command (hadoop jar
> GraphClean args) from the command line it works fine. I am using hadoop
> 0.19.0.
>
> Thanks,
>
> Razen
>
>
>
--
Best Regards, Edward J. Yoon @ NHN, corp.
[email protected]
http://blog.udanax.org