Hi Uma,

I suggest you do a 'mvn install -DskipTests' before running 'mvn 
eclipse:eclipse'.

Thanks,
Chuan

-----Original Message-----
From: Uma Maheswara Rao G [mailto:hadoop....@gmail.com] 
Sent: Friday, July 12, 2013 7:42 PM
To: common-...@hadoop.apache.org
Cc: hdfs-dev@hadoop.apache.org
Subject: Re: mvn eclipse:eclipse failure on windows

HI Chris,
  eclipse:eclipse works but still I am seeing UnsatisfiesLinkError.
Explicitly I pointed java.library.path to where hadoop.dll geneated. This
dll generated with my clean install command only.   My pc is 64 but and
also set Platform=x64 while building. But does not help.

Regards,
Uma






On Fri, Jul 12, 2013 at 11:45 PM, Chris Nauroth <cnaur...@hortonworks.com>wrote:

> Hi Uma,
>
> I just tried getting a fresh copy of trunk and running "mvn clean 
> install -DskipTests" followed by "mvn eclipse:eclipse -DskipTests".  
> Everything worked fine in my environment.  Are you still seeing the problem?
>
> The UnsatisfiedLinkError seems to indicate that your build couldn't 
> access hadoop.dll for JNI method implementations.  hadoop.dll gets 
> built as part of the hadoop-common sub-module.  Is it possible that 
> you didn't have a complete package build for that sub-module before 
> you started running the HDFS test?
>
> Chris Nauroth
> Hortonworks
> http://hortonworks.com/
>
>
>
> On Sun, Jul 7, 2013 at 9:08 AM, sure bhands <sure.bha...@gmail.com> wrote:
>
> > I would try cleaning hadoop-maven-plugin directory from maven 
> > repository
> to
> > rule out the stale version and then mv install followed by mvn 
> > eclipse:eclipse before digging in to it further.
> >
> > Thanks,
> > Surendra
> >
> >
> > On Sun, Jul 7, 2013 at 8:28 AM, Uma Maheswara Rao G <
> hadoop....@gmail.com
> > >wrote:
> >
> > > Hi,
> > >
> > > I am seeing this failure on windows while executing mvn 
> > > eclipse:eclipse command on trunk.
> > >
> > > See the following trace:
> > >
> > > [INFO]
> > >
> ----------------------------------------------------------------------
> --
> > > [ERROR] Failed to execute goal
> > > org.apache.maven.plugins:maven-eclipse-plugin:2.8
> > > :eclipse (default-cli) on project hadoop-common: Request to merge 
> > > when 'filterin g' is not identical. Original=resource 
> > > src/main/resources:
> > > output=target/classes
> > > , include=[], exclude=[common-version-info.properties|**/*.java],
> > > test=false, fi
> > > ltering=false, merging with=resource src/main/resources:
> > > output=target/classes,
> > > include=[common-version-info.properties], exclude=[**/*.java],
> > test=false,
> > > filte
> > > ring=true -> [Help 1]
> > > [ERROR]
> > > [ERROR] To see the full stack trace of the errors, re-run Maven 
> > > with
> the
> > -e
> > > swit
> > > ch.
> > > [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> > > [ERROR]
> > > [ERROR] For more information about the errors and possible 
> > > solutions, please rea d the following articles:
> > > [ERROR] [Help 1]
> > > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionE
> > > xception
> > > [ERROR]
> > > [ERROR] After correcting the problems, you can resume the build 
> > > with
> the
> > > command
> > >
> > > [ERROR]   mvn <goals> -rf :hadoop-common
> > > E:\Hadoop-Trunk>
> > >
> > > any idea for resolving it.
> > >
> > > With 'org.apache.maven.plugins:maven-eclipse-plugin:2.6:eclipse' 
> > > seems
> to
> > > be no failures but  I am seeing following exception while running
> tests.
> > > java.lang.UnsatisfiedLinkError:
> > >
> > >
> >
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/Stri
> ng;I)Z
> > >     at 
> > > org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native
> > > Method)
> > >     at
> > >
> org.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:42
> 3)
> > >     at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:952)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeS
> torage(Storage.java:451)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSIm
> age.java:282)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(F
> SImage.java:200)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSName
> system.java:696)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNam
> esystem.java:530)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNod
> e.java:401)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.ja
> va:435)
> > >     at
> > >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:6
> 07)
> > >     at
> > >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:5
> 92)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNod
> e.java:1172)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.ja
> va:895)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDF
> SCluster.java:786)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluste
> r.java:644)
> > >     at
> > > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:334)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.jav
> a:316)
> > >     at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.namenode.ha.TestHASafeMode.setupCluster(
> TestHASafeMode.java:87)
> > >
> > > Not sure what I missed here. Any idea what could be wrong here?
> > >
> > > Regards,
> > > Uma
> > >
> >
>

Reply via email to