HI Chris,
  eclipse:eclipse works but still I am seeing UnsatisfiesLinkError.
Explicitly I pointed java.library.path to where hadoop.dll geneated. This
dll generated with my clean install command only.   My pc is 64 but and
also set Platform=x64 while building. But does not help.

Regards,
Uma






On Fri, Jul 12, 2013 at 11:45 PM, Chris Nauroth <cnaur...@hortonworks.com>wrote:

> Hi Uma,
>
> I just tried getting a fresh copy of trunk and running "mvn clean install
> -DskipTests" followed by "mvn eclipse:eclipse -DskipTests".  Everything
> worked fine in my environment.  Are you still seeing the problem?
>
> The UnsatisfiedLinkError seems to indicate that your build couldn't access
> hadoop.dll for JNI method implementations.  hadoop.dll gets built as part
> of the hadoop-common sub-module.  Is it possible that you didn't have a
> complete package build for that sub-module before you started running the
> HDFS test?
>
> Chris Nauroth
> Hortonworks
> http://hortonworks.com/
>
>
>
> On Sun, Jul 7, 2013 at 9:08 AM, sure bhands <sure.bha...@gmail.com> wrote:
>
> > I would try cleaning hadoop-maven-plugin directory from maven repository
> to
> > rule out the stale version and then mv install followed by mvn
> > eclipse:eclipse before digging in to it further.
> >
> > Thanks,
> > Surendra
> >
> >
> > On Sun, Jul 7, 2013 at 8:28 AM, Uma Maheswara Rao G <
> hadoop....@gmail.com
> > >wrote:
> >
> > > Hi,
> > >
> > > I am seeing this failure on windows while executing mvn eclipse:eclipse
> > > command on trunk.
> > >
> > > See the following trace:
> > >
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > [ERROR] Failed to execute goal
> > > org.apache.maven.plugins:maven-eclipse-plugin:2.8
> > > :eclipse (default-cli) on project hadoop-common: Request to merge when
> > > 'filterin
> > > g' is not identical. Original=resource src/main/resources:
> > > output=target/classes
> > > , include=[], exclude=[common-version-info.properties|**/*.java],
> > > test=false, fi
> > > ltering=false, merging with=resource src/main/resources:
> > > output=target/classes,
> > > include=[common-version-info.properties], exclude=[**/*.java],
> > test=false,
> > > filte
> > > ring=true -> [Help 1]
> > > [ERROR]
> > > [ERROR] To see the full stack trace of the errors, re-run Maven with
> the
> > -e
> > > swit
> > > ch.
> > > [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> > > [ERROR]
> > > [ERROR] For more information about the errors and possible solutions,
> > > please rea
> > > d the following articles:
> > > [ERROR] [Help 1]
> > > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionE
> > > xception
> > > [ERROR]
> > > [ERROR] After correcting the problems, you can resume the build with
> the
> > > command
> > >
> > > [ERROR]   mvn <goals> -rf :hadoop-common
> > > E:\Hadoop-Trunk>
> > >
> > > any idea for resolving it.
> > >
> > > With 'org.apache.maven.plugins:maven-eclipse-plugin:2.6:eclipse' seems
> to
> > > be no failures but  I am seeing following exception while running
> tests.
> > > java.lang.UnsatisfiedLinkError:
> > >
> > >
> >
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
> > >     at org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native
> > > Method)
> > >     at
> > >
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:423)
> > >     at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:952)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:451)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:282)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:200)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:696)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:530)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:401)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:435)
> > >     at
> > >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:607)
> > >     at
> > >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:592)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1172)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:895)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:786)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:644)
> > >     at
> > > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:334)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:316)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.ha.TestHASafeMode.setupCluster(TestHASafeMode.java:87)
> > >
> > > Not sure what I missed here. Any idea what could be wrong here?
> > >
> > > Regards,
> > > Uma
> > >
> >
>

Reply via email to