[ https://issues.apache.org/jira/browse/SQOOP-3043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15635821#comment-15635821 ]
Ramesh B commented on SQOOP-3043: --------------------------------- This is due to following code {code} private void removeTempLogs(Path tablePath) throws IOException { FileSystem fs = FileSystem.get(configuration); Path logsPath = new Path(tablePath, "_logs"); if (fs.exists(logsPath)) {... {code} I have provided pull request here https://github.com/apache/sqoop/pull/29/files > Sqoop HiveImport fails with Wrong FS while removing the _logs > -------------------------------------------------------------- > > Key: SQOOP-3043 > URL: https://issues.apache.org/jira/browse/SQOOP-3043 > Project: Sqoop > Issue Type: Bug > Components: hive-integration > Reporter: Ramesh B > > With s3:// as --target-dir and --hive-import provided sqoop fails with > {code}ERROR tool.ImportTool: Imported Failed: Wrong FS: > s3a://dataplatform/sqoop/target/user/_logs, expected: hdfs://nn1 > {code} > This is due to removeTempLogs method in HiveImport.java which is expecting > hdfs as the path. -- This message was sent by Atlassian JIRA (v6.3.4#6332)