Re: s3 bucket access/read file

2017-05-17 Thread Steve Loughran
On 17 May 2017, at 00:10, jazzed mailto:crackshotm...@gmail.com>> wrote: How did you solve the problem with V4? which v4 problem? Authentication? you need to declare the explicit s3a endpoint via fs.s3a.endpoint , otherwise you get a generic "bad auth" message which is not a good place to st

Re: s3 bucket access/read file

2017-05-16 Thread jazzed
How did you solve the problem with V4? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/s3-bucket-access-read-file-tp23536p28688.html Sent from the Apache Spark User List mailing list archive at Nabble.com. -

Re: s3 bucket access/read file

2015-07-01 Thread Aaron Davidson
I think 2.6 failed to abruptly close streams that weren't fully read, which we observed as a huge performance hit. We had to backport the 2.7 improvements before being able to use it.

Re: s3 bucket access/read file

2015-07-01 Thread Steve Loughran
s3a uses amazon's own libraries; it's tested against frankfurt too. you have to view s3a support in Hadoop 2.6 as beta-release: it works, with some issues. Hadoop 2.7.0+ has it all working now, though are left with the task of getting hadoop-aws and the amazon JAR onto your classpath via the --j

Re: s3 bucket access/read file

2015-06-30 Thread Aaron Davidson
Should be able to use s3a (on new hadoop versions), I believe that will try or at least has a setting for v4 On Tue, Jun 30, 2015 at 8:31 PM, Exie wrote: > Not sure if this helps, but the options I set are slightly different: > > val hadoopConf=sc.hadoopConfiguration > hadoopConf.set("fs.s3n.aws

Re: s3 bucket access/read file

2015-06-30 Thread Exie
Not sure if this helps, but the options I set are slightly different: val hadoopConf=sc.hadoopConfiguration hadoopConf.set("fs.s3n.awsAccessKeyId","key") hadoopConf.set("fs.s3n.awsSecretAccessKey","secret") Try setting them to s3n as opposed to just s3 Good luck! -- View this message in conte

Re: s3 bucket access/read file

2015-06-30 Thread didi
We finally managed to find the problem, the s3 files were located in Frankfurt which only supports the *v4* signature *Surprising* is the fact that the spark core library method textfile does not support that!! -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.co

Re: s3 bucket access/read file

2015-06-30 Thread Akhil Das
Try this way: val data = sc.textFile("s3n://ACCESS_KEY:SECRET_KEY@mybucket/temp/") Thanks Best Regards On Mon, Jun 29, 2015 at 11:59 PM, didi wrote: > Hi > > *Cant read text file from s3 to create RDD > * > > after setting the configuration > val hadoopConf=sparkContext.hadoopConfiguration; >

Re: s3 bucket access/read file

2015-06-29 Thread spark user
Pls check your ACL properties. On Monday, June 29, 2015 11:29 AM, didi wrote: Hi *Cant read text file from s3 to create RDD * after setting the configuration val hadoopConf=sparkContext.hadoopConfiguration; hadoopConf.set("fs.s3.impl", "org.apache.hadoop.fs.s3native.NativeS3FileSys