Sahil Takiar created HIVE-15502: ----------------------------------- Summary: CTAS on S3 is broken with credentials exception Key: HIVE-15502 URL: https://issues.apache.org/jira/browse/HIVE-15502 Project: Hive Issue Type: Bug Components: Hive Reporter: Sahil Takiar Assignee: Sahil Takiar
Simple CTAS queries that read from S3, and write to the local fs throw the following exception: {code} com.amazonaws.AmazonClientException: Unable to load AWS credentials from any provider in the chain at com.amazonaws.auth.AWSCredentialsProviderChain.getCredentials(AWSCredentialsProviderChain.java:117) at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3521) at com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031) at com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994) at org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2653) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.hadoop.hive.ql.exec.Utilities.isEmptyPath(Utilities.java:2308) at org.apache.hadoop.hive.ql.exec.Utilities.isEmptyPath(Utilities.java:2304) at org.apache.hadoop.hive.ql.exec.Utilities.getInputPaths(Utilities.java:3013) at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:342) at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:151) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2168) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1824) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1511) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1222) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1212) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:400) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:777) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:715) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:642) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136) Job Submission failed with exception 'com.amazonaws.AmazonClientException(Unable to load AWS credentials from any provider in the chain)' {code} Seems to only happen when trying to connect to S3 from map tasks. My {{hive-site.xml}} has the following entries: {code} <configuration> <property> <name>mapreduce.framework.name</name> <value>local</value> </property> <property> <name>mapred.job.tracker</name> <value>local</value> </property> <property> <name>fs.default.name</name> <value>file:///</value> </property> <property> <name>fs.s3a.access.key</name> <value>[ACCESS-KEY]</value> </property> <property> <name>fs.s3a.secret.key</name> <value>[SECRET-KEY]</value> </property> </configuration> {code} I've also noticed that now I need to copy the AWS S3 SDK jars into the lib folder before running Hive locally. -- This message was sent by Atlassian JIRA (v6.3.4#6332)