Hi raj,
If you have all data on NFS mounted disk, meaning on single machine, then
your upload will be limited by network bandwidth. You can try running dfs
-put in multiple parallel threads for distinct data sets, you might be able
to utilise network bandwidth to its maximum(take care not to have
[
https://issues.apache.org/jira/browse/HADOOP-7989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jai Kumar Singh resolved HADOOP-7989.
-
Resolution: Not A Problem
It was a setting problem. (by whirr) ulimit was not set to unl
> If it is divided up into several files and you can mount your NFS
directory on each of the datanodes.
Just curious, how will this help.
Praveen
On Wed, Jan 25, 2012 at 12:39 AM, Robert Evans wrote:
> If it is divided up into several files and you can mount your NFS
> directory on each of the
Unsure on what could cause that but I'd first check if Cygwin's bin/
items are on my Windows PATH.
FWIW, if you are stepping into development, and your development isn't
windows-specific, best to use any modern Linux distro as your
environment instead. Ubuntu works great for starters. I prefer usi
HA: the FailoverController should check the standby is ready before failing over
Key: HADOOP-7991
URL: https://issues.apache.org/jira/browse/HADOOP-7991
Project: Hadoop C
[
https://issues.apache.org/jira/browse/HADOOP-7571?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Eli Collins resolved HADOOP-7571.
-
Resolution: Fixed
Looks like this was fixed.
> hadoop-config.sh needs to be upd
Hello,
I am new to hadoop as well. I just followed the instruction,got the source
code and tried to build it by maven .
during the mvn install, build failed and got an error which says :
[ERROR] Failed to execute Goal org.codebus.mojo: Exec-Maven-plugin: 1.2:
Exec on project hadoop-yarn-common
TFile to allow Snappy compression
-
Key: HADOOP-7990
URL: https://issues.apache.org/jira/browse/HADOOP-7990
Project: Hadoop Common
Issue Type: Improvement
Reporter: Prashant Kommireddi
TFile curr
If it is divided up into several files and you can mount your NFS directory on
each of the datanodes, you could possibly use distcp to do it. I have never
tried using distcp for this, but it should work. Or you can write your own
streaming Map/Reduce script that does more or less the same thin
You can currently get CPU & memory stats for each task and aggregated stats per
job via MapReduce Counters.
Arun
On Jan 24, 2012, at 7:27 AM, arindam choudhury wrote:
> Hi,
>
> How can I get CPU, memory, network and disk usage statistics of a hadoop
> job?
>
> Thanks,
> Arindam
Hi,
How can I get CPU, memory, network and disk usage statistics of a hadoop
job?
Thanks,
Arindam
Ashok,
Following http://wiki.apache.org/hadoop/HowToContribute should get you
started at development. Let us know if you have any further, specific
questions once you get ahead.
On Tue, Jan 24, 2012 at 7:30 PM, ashok chandrasekar
wrote:
> Hi,
> I am new to hadoop. I want to make a few changes to
Hi,
I am new to hadoop. I want to make a few changes to the hadoop source code
for my college project. How can I do this? How to compile and test the
modified code? What tools are needed to perform this? Your reply will be of
great help to me.
Thanks in advance.
13 matches
Mail list logo