I would try cleaning hadoop-maven-plugin directory from maven repository to
rule out the stale version and then mv install followed by mvn
eclipse:eclipse before digging in to it further.

Thanks,
Surendra


On Sun, Jul 7, 2013 at 8:28 AM, Uma Maheswara Rao G <hadoop....@gmail.com>wrote:

> Hi,
>
> I am seeing this failure on windows while executing mvn eclipse:eclipse
> command on trunk.
>
> See the following trace:
>
> [INFO]
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal
> org.apache.maven.plugins:maven-eclipse-plugin:2.8
> :eclipse (default-cli) on project hadoop-common: Request to merge when
> 'filterin
> g' is not identical. Original=resource src/main/resources:
> output=target/classes
> , include=[], exclude=[common-version-info.properties|**/*.java],
> test=false, fi
> ltering=false, merging with=resource src/main/resources:
> output=target/classes,
> include=[common-version-info.properties], exclude=[**/*.java], test=false,
> filte
> ring=true -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e
> swit
> ch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions,
> please rea
> d the following articles:
> [ERROR] [Help 1]
> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionE
> xception
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the
> command
>
> [ERROR]   mvn <goals> -rf :hadoop-common
> E:\Hadoop-Trunk>
>
> any idea for resolving it.
>
> With 'org.apache.maven.plugins:maven-eclipse-plugin:2.6:eclipse' seems to
> be no failures but  I am seeing following exception while running tests.
> java.lang.UnsatisfiedLinkError:
>
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
>     at org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native
> Method)
>     at
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:423)
>     at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:952)
>     at
>
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:451)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:282)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:200)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:696)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:530)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:401)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:435)
>     at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:607)
>     at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:592)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1172)
>     at
>
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:895)
>     at
>
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:786)
>     at
>
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:644)
>     at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:334)
>     at
>
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:316)
>     at
>
> org.apache.hadoop.hdfs.server.namenode.ha.TestHASafeMode.setupCluster(TestHASafeMode.java:87)
>
> Not sure what I missed here. Any idea what could be wrong here?
>
> Regards,
> Uma
>

Reply via email to