I played on Hadoop MapReduce on Riak CS, and it actually worked with the latest 1.5 beta package. Hadoop relies S3 connectivity on jets3t, so if MapR uses vanilla jets3t it will work. I believe so because MapR works on EMR (which usually extracts data from S3).
Technically, you can add several options about S3 endpoints to connect other S3-compatible cloud storages into jets3t.properties, which are mainly "s3service.s3-endpoint" and "s3service.s3-endpoint-http(s)-port". I put the properties file into hadoop conf directory and it worked. Maybe there is a config-loading in MapR, too. [1] In this case, you should properly configure your CS use your domain by cs_root_host in app.config. [2] If your Riak CS is not configured with your own domain, you can also configure MapReduce to use proxy setting like this: httpclient.proxy-host=localhost httpclient.proxy-port=8080 I usually use this configuration when I play locally. Put them into jets3t.properties. Note that 1.4.x CS won't work properly if the output file is on CS again - it doesn't have copy API used in the final file copy after reduce. We have 1.5 pre-release package internally and testing. Sooner or later it will be released. [1] https://jets3t.s3.amazonaws.com/toolkit/configuration.html [2] http://docs.basho.com/riakcs/latest/cookbooks/configuration/Configuring-Riak-CS/ On Fri, Aug 1, 2014 at 4:08 AM, John Daily <jda...@basho.com> wrote: > This blog post on configuring S3 clients to work with CS may be useful: > http://basho.com/riak-cs-proxy-vs-direct-configuration/ > > Sent from my iPhone > > On Jul 31, 2014, at 2:53 PM, Andrew Stone <ast...@basho.com> wrote: > > Hi Charles, > > AFAIK we haven't ever tested Riak Cs with the MapR connector. However, if > MapR works with S3, you should just have to change the IP to point to a load > balancer in front of your local Riak CS cluster. I'm unaware of how to > change that setting in MapR though. It seems like a question for them and > not Basho. > > -Andrew > > > On Wed, Jul 30, 2014 at 5:16 PM, Charles Shah <find.chuck...@gmail.com> > wrote: >> >> Hi, >> >> I would like to use MapR with Riak CS for hadoop map reduce jobs. My code >> is currently referring to objects using s3n:// urls. >> I'd like to be able to have the hadoop code on MapR point to the Riak CS >> cluster using the s3 url. >> Is there a proxy or hostname setting in hadoop to be able to route the s3 >> url to the riak cs cluster ? >> >> Thanks >> >> >> _______________________________________________ >> riak-users mailing list >> riak-users@lists.basho.com >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >> > > _______________________________________________ > riak-users mailing list > riak-users@lists.basho.com > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com > > > _______________________________________________ > riak-users mailing list > riak-users@lists.basho.com > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com > -- Kota UENISHI / @kuenishi Basho Japan KK _______________________________________________ riak-users mailing list riak-users@lists.basho.com http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com