Is this a bug?
I’ve tried the same configuration files with hadoop 2.6.0, hbase 1.0.0 and it 
is working. But when I am using the same configuration files, and doing the 
same thing with hadoop 2.5.1 and hbase 0.98.7-hadoop2 I have this errors

hbase-0.98.7-hadoop2/bin/hbase org.apache.hadoop.hbase.snapshot.ExportSnapshot 
-snapshot mysnap -copy-to hdfs://198.58.88.31:9000/hbase/test
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/home/vagrant/hbase-0.98.7-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/home/vagrant/hadoop-2.5.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
2015-03-17 16:38:08,815 INFO  [main] snapshot.ExportSnapshot: Copy Snapshot 
Manifest
2015-03-17 16:38:09,235 INFO  [main] Configuration.deprecation: session.id is 
deprecated. Instead, use dfs.metrics.session-id
2015-03-17 16:38:09,236 INFO  [main] jvm.JvmMetrics: Initializing JVM Metrics 
with processName=JobTracker, sessionId=
2015-03-17 16:38:09,334 INFO  [main] mapreduce.JobSubmitter: Cleaning up the 
staging area 
file:/home/vagrant/hadoop-2.5.1/hadoop-datastore/mapred/staging/vagrant1812768224/.staging/job_local1812768224_0001
2015-03-17 16:38:09,335 ERROR [main] snapshot.ExportSnapshot: Snapshot export 
failed
java.io.FileNotFoundException: File does not exist: 
hdfs://namenode:9000/home/vagrant/hbase-0.98.7-hadoop2/lib/hadoop-mapreduce-client-core-2.5.1.jar
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1072)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1064)
        at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1064)
        at 
org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:288)
        at 
org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:224)
        at 
org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestamps(ClientDistributedCacheManager.java:93)
        at 
org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestampsAndCacheVisibilities(ClientDistributedCacheManager.java:57)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:265)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:301)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:389)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1303)
        at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot.runCopyJob(ExportSnapshot.java:768)
        at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot.run(ExportSnapshot.java:925)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
        at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot.innerMain(ExportSnapshot.java:991)
        at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot.main(ExportSnapshot.java:995)

Any ideas?
 
> On 17 Mar 2015, at 12:38, Akmal Abbasov <[email protected]> wrote:
> 
> Hi, I have 2 clusters running HBase, and I want to export a snapshot from 
> cluster A to cluster B.
> When I am doing exportSnapshot I am getting java.io.FileNotFoundException, 
> because it is searching for a jar file in hdfs, not in my local storage.
> Any ideas how it could be solved?
> 
> Here is an output:
> 2015-03-17 11:30:26,310 INFO  [main] Configuration.deprecation: session.id is 
> deprecated. Instead, use dfs.metrics.session-id
> 2015-03-17 11:30:26,312 INFO  [main] jvm.JvmMetrics: Initializing JVM Metrics 
> with processName=JobTracker, sessionId=
> 2015-03-17 11:30:27,383 INFO  [main] mapreduce.JobSubmitter: Cleaning up the 
> staging area 
> file:/tmp/hadoop-hadoop/mapred/staging/hadoop806729561/.staging/job_local806729561_0001
> 2015-03-17 11:30:27,387 ERROR [main] snapshot.ExportSnapshot: Snapshot export 
> failed
> java.io.FileNotFoundException: File does not exist: 
> hdfs://hiveprodeuw1/opt/hadoop/hbase-0.98.7-hadoop2/lib/hadoop-common-2.5.1.jar
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1072)
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1064)
>       at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1064)
>       at 
> org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:288)
>       at 
> org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:224)
>       at 
> org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestamps(ClientDistributedCacheManager.java:93)
>       at 
> org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestampsAndCacheVisibilities(ClientDistributedCacheManager.java:57)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:265)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:301)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:389)
>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
>       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
>       at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1303)
>       at 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot.runCopyJob(ExportSnapshot.java:768)
>       at 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot.run(ExportSnapshot.java:925)
>       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>       at 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot.innerMain(ExportSnapshot.java:991)
>       at 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot.main(ExportSnapshot.java:995)

Reply via email to