Hi Uma,

I just tried getting a fresh copy of trunk and running "mvn clean install
-DskipTests" followed by "mvn eclipse:eclipse -DskipTests".  Everything
worked fine in my environment.  Are you still seeing the problem?

The UnsatisfiedLinkError seems to indicate that your build couldn't access
hadoop.dll for JNI method implementations.  hadoop.dll gets built as part
of the hadoop-common sub-module.  Is it possible that you didn't have a
complete package build for that sub-module before you started running the
HDFS test?

Chris Nauroth
Hortonworks
http://hortonworks.com/



On Sun, Jul 7, 2013 at 9:08 AM, sure bhands <sure.bha...@gmail.com> wrote:

> I would try cleaning hadoop-maven-plugin directory from maven repository to
> rule out the stale version and then mv install followed by mvn
> eclipse:eclipse before digging in to it further.
>
> Thanks,
> Surendra
>
>
> On Sun, Jul 7, 2013 at 8:28 AM, Uma Maheswara Rao G <hadoop....@gmail.com
> >wrote:
>
> > Hi,
> >
> > I am seeing this failure on windows while executing mvn eclipse:eclipse
> > command on trunk.
> >
> > See the following trace:
> >
> > [INFO]
> > ------------------------------------------------------------------------
> > [ERROR] Failed to execute goal
> > org.apache.maven.plugins:maven-eclipse-plugin:2.8
> > :eclipse (default-cli) on project hadoop-common: Request to merge when
> > 'filterin
> > g' is not identical. Original=resource src/main/resources:
> > output=target/classes
> > , include=[], exclude=[common-version-info.properties|**/*.java],
> > test=false, fi
> > ltering=false, merging with=resource src/main/resources:
> > output=target/classes,
> > include=[common-version-info.properties], exclude=[**/*.java],
> test=false,
> > filte
> > ring=true -> [Help 1]
> > [ERROR]
> > [ERROR] To see the full stack trace of the errors, re-run Maven with the
> -e
> > swit
> > ch.
> > [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> > [ERROR]
> > [ERROR] For more information about the errors and possible solutions,
> > please rea
> > d the following articles:
> > [ERROR] [Help 1]
> > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionE
> > xception
> > [ERROR]
> > [ERROR] After correcting the problems, you can resume the build with the
> > command
> >
> > [ERROR]   mvn <goals> -rf :hadoop-common
> > E:\Hadoop-Trunk>
> >
> > any idea for resolving it.
> >
> > With 'org.apache.maven.plugins:maven-eclipse-plugin:2.6:eclipse' seems to
> > be no failures but  I am seeing following exception while running tests.
> > java.lang.UnsatisfiedLinkError:
> >
> >
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
> >     at org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native
> > Method)
> >     at
> > org.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:423)
> >     at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:952)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:451)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:282)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:200)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:696)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:530)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:401)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:435)
> >     at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:607)
> >     at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:592)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1172)
> >     at
> >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:895)
> >     at
> >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:786)
> >     at
> >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:644)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:334)
> >     at
> >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:316)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.ha.TestHASafeMode.setupCluster(TestHASafeMode.java:87)
> >
> > Not sure what I missed here. Any idea what could be wrong here?
> >
> > Regards,
> > Uma
> >
>

Reply via email to