I am a big data developer and for past some time I have been using apache spark of cloudera distribution. I have successfully written all my codes on Cloudera-Spark but when I am trying my hands on MapR distribution of Spark then I am getting the following error.
=============================================== java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at com.mapr.fs.ShimLoader.loadNativeLibrary(ShimLoader.java:308) at com.mapr.fs.ShimLoader.load(ShimLoader.java:197) at org.apache.hadoop.conf.CoreDefaultProperties.<clinit>(CoreDefaultProperties.java:54) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Unknown Source) at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1822) at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2037) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2238) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2190) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2107) at org.apache.hadoop.conf.Configuration.set(Configuration.java:967) at org.apache.hadoop.conf.Configuration.set(Configuration.java:941) at org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:102) at org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$.<init>(SparkHadoopUtil.scala:202) at org.apache.spark.deploy.SparkHadoopUtil$.<clinit>(SparkHadoopUtil.scala) at org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1784) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:105) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:180) at org.apache.spark.SparkEnv$.create(SparkEnv.scala:292) at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:159) at org.apache.spark.SparkContext.<init>(SparkContext.scala:232) at org.apache.spark.SparkContext.<init>(SparkContext.scala:136) at org.apache.spark.SparkContext.<init>(SparkContext.scala:151) at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:67) at word.count.WordCount.main(WordCount.java:28) Caused by: java.lang.UnsatisfiedLinkError: no MapRClient in java.library.path at java.lang.ClassLoader.loadLibrary(Unknown Source) at java.lang.Runtime.loadLibrary0(Unknown Source) at java.lang.System.loadLibrary(Unknown Source) at com.mapr.fs.shim.LibraryLoader.loadLibrary(LibraryLoader.java:41) ... 30 more ==========Unable to find library in jar due to exception. ============== java.lang.RuntimeException: no native library is found for os.name=Windows and os.arch=x86_64 at com.mapr.fs.ShimLoader.findNativeLibrary(ShimLoader.java:496) at com.mapr.fs.ShimLoader.loadNativeLibrary(ShimLoader.java:318) at com.mapr.fs.ShimLoader.load(ShimLoader.java:197) at org.apache.hadoop.conf.CoreDefaultProperties.<clinit>(CoreDefaultProperties.java:54) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Unknown Source) at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1822) at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2037) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2238) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2190) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2107) at org.apache.hadoop.conf.Configuration.set(Configuration.java:967) at org.apache.hadoop.conf.Configuration.set(Configuration.java:941) at org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:102) at org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$.<init>(SparkHadoopUtil.scala:202) at org.apache.spark.deploy.SparkHadoopUtil$.<clinit>(SparkHadoopUtil.scala) at org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1784) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:105) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:180) at org.apache.spark.SparkEnv$.create(SparkEnv.scala:292) at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:159) at org.apache.spark.SparkContext.<init>(SparkContext.scala:232) at org.apache.spark.SparkContext.<init>(SparkContext.scala:136) at org.apache.spark.SparkContext.<init>(SparkContext.scala:151) at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:67) at word.count.WordCount.main(WordCount.java:28) java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at com.mapr.fs.ShimLoader.loadNativeLibrary(ShimLoader.java:308) at com.mapr.fs.ShimLoader.load(ShimLoader.java:197) at org.apache.hadoop.conf.CoreDefaultProperties.<clinit>(CoreDefaultProperties.java:54) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Unknown Source) at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1822) at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2037) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2238) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2190) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2107) at org.apache.hadoop.conf.Configuration.set(Configuration.java:967) at org.apache.hadoop.conf.Configuration.set(Configuration.java:941) at org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:102) at org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$.<init>(SparkHadoopUtil.scala:202) at org.apache.spark.deploy.SparkHadoopUtil$.<clinit>(SparkHadoopUtil.scala) at org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1784) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:105) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:180) at org.apache.spark.SparkEnv$.create(SparkEnv.scala:292) at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:159) at org.apache.spark.SparkContext.<init>(SparkContext.scala:232) at org.apache.spark.SparkContext.<init>(SparkContext.scala:136) at org.apache.spark.SparkContext.<init>(SparkContext.scala:151) at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:67) at word.count.WordCount.main(WordCount.java:28) Caused by: java.lang.UnsatisfiedLinkError: no MapRClient in java.library.path at java.lang.ClassLoader.loadLibrary(Unknown Source) at java.lang.Runtime.loadLibrary0(Unknown Source) at java.lang.System.loadLibrary(Unknown Source) at com.mapr.fs.shim.LibraryLoader.loadLibrary(LibraryLoader.java:41) ... 30 more Exception in thread "main" java.lang.ExceptionInInitializerError at com.mapr.fs.ShimLoader.load(ShimLoader.java:214) at org.apache.hadoop.conf.CoreDefaultProperties.<clinit>(CoreDefaultProperties.java:54) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Unknown Source) at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1822) at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2037) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2238) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2190) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2107) at org.apache.hadoop.conf.Configuration.set(Configuration.java:967) at org.apache.hadoop.conf.Configuration.set(Configuration.java:941) at org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:102) at org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$.<init>(SparkHadoopUtil.scala:202) at org.apache.spark.deploy.SparkHadoopUtil$.<clinit>(SparkHadoopUtil.scala) at org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1784) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:105) at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:180) at org.apache.spark.SparkEnv$.create(SparkEnv.scala:292) at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:159) at org.apache.spark.SparkContext.<init>(SparkContext.scala:232) at org.apache.spark.SparkContext.<init>(SparkContext.scala:136) at org.apache.spark.SparkContext.<init>(SparkContext.scala:151) at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:67) at word.count.WordCount.main(WordCount.java:28) Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at com.mapr.fs.ShimLoader.loadNativeLibrary(ShimLoader.java:308) at com.mapr.fs.ShimLoader.load(ShimLoader.java:197) ... 24 more Caused by: java.lang.UnsatisfiedLinkError: no MapRClient in java.library.path at java.lang.ClassLoader.loadLibrary(Unknown Source) at java.lang.Runtime.loadLibrary0(Unknown Source) at java.lang.System.loadLibrary(Unknown Source) at com.mapr.fs.shim.LibraryLoader.loadLibrary(LibraryLoader.java:41) ... 30 more ==================================== I have written a simple word count code. *When I am removing MapR-Spark jar and replacing it with Cloudera-Spark jar then the same code is executing perfectly*. Kindly help. -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/MapR-distribution-Spark-throwing-no-MapRClient-in-java-library-path-error-tp9922.html Sent from the Apache Spark Developers List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org