Hey Nikaash, It was mistake from my end I ran only start-master.sh command due to which there were no workers. :( Later I ran comman start-all.sh and it worked for me. Thanks for the help.
On Thu, Apr 14, 2016 at 4:53 PM, Rohit Jain <[email protected]> wrote: > Hey Nikaash, > thanks for the help. command is running but its not aasigning any worker > to job. IT keep keeps showwing this following message. > Initial job has not accepted any resources; check your cluster UI to > ensure that workers are registered and have sufficient resources > > On Thu, Apr 14, 2016 at 3:48 PM, Nikaash Puri <[email protected]> > wrote: > >> my guess is that its not initialising the Configuration correctly since >> the spark master is by default local. My guess is that to read from HDFS >> you’d have to set up a Spark cluster and specify the master URL using the >> —master flag. >> >> Else, set up a Spark on YARN deployment using CDH. Then specify the >> master as —yarn-cluster. That should also work. >> > On 14-Apr-2016, at 3:43 PM, Rohit Jain <[email protected]> wrote: >> > >> > Hello Nikaash >> > >> > I did this. >> > >> > /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/bin/mahout >> > spark-itemsimilarity -i hdfs://localhost:9000/spdemo.txt -o >> > hdfs://localhost:9000/spout/ --filter1 like -fc 1 -ic 2 >> > -D:spark.io.compression.codec=lzf >> > >> > but it throws this exception >> > >> > Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS: >> > hdfs://localhost:9000/spdemo1.txt, expected: file:/// >> > >> > On Thu, Apr 14, 2016 at 3:41 PM, Nikaash Puri <[email protected]> >> wrote: >> > >> >> Hi, >> >> >> >> You’d need to specify the complete path along with HDFS URI. for eg. >> >> hdfs://namenode:<port>/path/to/file. >> >> >> >> - Nikaash >> >>> On 14-Apr-2016, at 3:39 PM, Rohit Jain <[email protected]> >> wrote: >> >>> >> >>> Hey Nikaash, >> >>> >> >>> I ran this command but its working local files not for hadoop input >> file. >> >>> How to make it work on hadoop uri? >> >>> >> >>> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/bin/mahout >> >>> spark-itemsimilarity -i spdemo.txt -o /spout --filter1 like -fc 1 -ic >> 2 >> >>> -D:spark.io.compression.codec=lzf >> >>> >> >>> On Thu, Apr 14, 2016 at 3:19 PM, Nikaash Puri <[email protected]> >> >> wrote: >> >>> >> >>>> Cool. Even if you have problems setting up the library, just change >> the >> >>>> compression codec from snappy to lzf >> (-D:spark.io.compression.codec=lzf) >> >>>> like in the command. It might hurt performance, but it’ll get it to >> work >> >>>> >> >>>> - Nikaash >> >>>>> On 14-Apr-2016, at 3:17 PM, Rohit Jain <[email protected]> >> wrote: >> >>>>> >> >>>>> yes. i am downloading the that library and putting it java/lib/ >> path. >> >>>>> Anythingelse need to be done? >> >>>>> >> >>>>> On Thu, Apr 14, 2016 at 3:08 PM, Nikaash Puri < >> [email protected]> >> >>>> wrote: >> >>>>> >> >>>>>> Hi Rohit, >> >>>>>> >> >>>>>> It’s a snappy library issue. Are you running on Mac? >> >>>>>> Anyhow, run it like this: >> >>>>>> ./bin/mahout spark-itemsimilarity --input <in> --output <out> >> >>>>>> -D:spark.io.compression.codec=lzf >> >>>>>> >> >>>>>> Nikaash >> >>>>>> >> >>>>>>> On 14-Apr-2016, at 12:25 PM, Rohit Jain <[email protected]> >> >>>> wrote: >> >>>>>>> >> >>>>>>> Hello everyone, >> >>>>>>> I am new to mahout. I am trying to run this command >> >>>>>>> >> >>>>>>> /Documents/apache-mahout-distribution-0.11.2/bin/mahout >> >>>>>>> spark-itemsimilarity -i spdemo1.txt -o /spout --filter1 like -fc 1 >> >> -ic >> >>>> 2 >> >>>>>>> >> >>>>>>> before running the command i uploaded file to hdfs >> >>>>>>> >> >>>>>>> bin/hadoop dfs -put spdemo.txt /spdemo1.txt >> >>>>>>> >> >>>>>>> but i am geting this long list of error.Do I need to install spark >> >> too? >> >>>>>>> >> >>>>>>> >> >>>>>>> SLF4J: Class path contains multiple SLF4J bindings. >> >>>>>>> >> >>>>>>> SLF4J: Found binding in >> >>>>>>> >> >>>>>> >> >>>> >> >> >> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-examples-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >>>>>>> >> >>>>>>> SLF4J: Found binding in >> >>>>>>> >> >>>>>> >> >>>> >> >> >> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-mr-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >>>>>>> >> >>>>>>> SLF4J: Found binding in >> >>>>>>> >> >>>>>> >> >>>> >> >> >> [jar:file:/Users/rohitjain/Documents/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >>>>>>> >> >>>>>>> SLF4J: Found binding in >> >>>>>>> >> >>>>>> >> >>>> >> >> >> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >>>>>>> >> >>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for >> an >> >>>>>>> explanation. >> >>>>>>> >> >>>>>>> SLF4J: Actual binding is of type >> [org.slf4j.impl.Log4jLoggerFactory] >> >>>>>>> >> >>>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 INFO SparkContext: Running Spark version 1.6.1 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 WARN NativeCodeLoader: Unable to load >> native-hadoop >> >>>>>>> library for your platform... using builtin-java classes where >> >>>> applicable >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing view acls to: >> >>>> rohitjain >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing modify acls to: >> >>>>>> rohitjain >> >>>>>>> >> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: SecurityManager: >> >> authentication >> >>>>>>> disabled; ui acls disabled; users with view permissions: >> >>>> Set(rohitjain); >> >>>>>>> users with modify permissions: Set(rohitjain) >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service >> >>>> 'sparkDriver' >> >>>>>> on >> >>>>>>> port 64062. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO Slf4jLogger: Slf4jLogger started >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO Remoting: Starting remoting >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO Remoting: Remoting started; listening on >> >>>> addresses >> >>>>>>> :[akka.tcp://[email protected]:64063] >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service >> >>>>>>> 'sparkDriverActorSystem' on port 64063. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering MapOutputTracker >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering BlockManagerMaster >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO DiskBlockManager: Created local directory >> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/blockmgr-6cdfd6f1-56c8-445c-95e0-70305b0f5f5a >> >>>>>>> >> >>>>>>> 16/04/14 10:53:11 INFO MemoryStore: MemoryStore started with >> capacity >> >>>>>> 2.4 GB >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkEnv: Registering >> OutputCommitCoordinator >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started >> >>>>>>> [email protected]:4040 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service >> 'SparkUI' >> >> on >> >>>>>>> port 4040. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkUI: Started SparkUI at >> >>>>>> http://192.168.1.122:4040 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO HttpFileServer: HTTP File server directory >> is >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO HttpServer: Starting HTTP Server >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started >> >>>>>>> [email protected]:64064 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service 'HTTP >> file >> >>>>>>> server' on port 64064. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-hdfs-0.11.2.jar >> >>>>>>> at http://192.168.1.122:64064/jars/mahout-hdfs-0.11.2.jar with >> >>>> timestamp >> >>>>>>> 1460611392248 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-0.11.2.jar >> >>>>>>> at http://192.168.1.122:64064/jars/mahout-math-0.11.2.jar with >> >>>> timestamp >> >>>>>>> 1460611392253 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-scala_2.10-0.11.2.jar >> >>>>>>> at >> http://192.168.1.122:64064/jars/mahout-math-scala_2.10-0.11.2.jar >> >>>>>> with >> >>>>>>> timestamp 1460611392255 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2-dependency-reduced.jar >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2-dependency-reduced.jar >> >>>>>>> with timestamp 1460611392303 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2.jar >> >>>>>>> at http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2.jar >> with >> >>>>>>> timestamp 1460611392304 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> may be removed in the future. Please use >> spark.kryoserializer.buffer >> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was >> >>>>>>> previously specified as '0.064'. Fractional values are no longer >> >>>>>> accepted. >> >>>>>>> To specify the equivalent now, one may use '64k'. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key >> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark >> 1.4 >> >>>> and >> >>>>>>> and may be removed in the future. Please use the new key >> >>>>>>> 'spark.kryoserializer.buffer' instead. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Executor: Starting executor ID driver on >> host >> >>>>>>> localhost >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service >> >>>>>>> 'org.apache.spark.network.netty.NettyBlockTransferService' on port >> >>>> 64065. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO NettyBlockTransferService: Server created >> on >> >>>> 64065 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Trying to register >> >>>>>> BlockManager >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMasterEndpoint: Registering >> block >> >>>>>>> manager localhost:64065 with 2.4 GB RAM, BlockManagerId(driver, >> >>>>>> localhost, >> >>>>>>> 64065) >> >>>>>>> >> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Registered BlockManager >> >>>>>>> >> >>>>>>> java.lang.reflect.InvocationTargetException >> >>>>>>> >> >>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> >>>>>>> >> >>>>>>> at java.lang.reflect.Method.invoke(Method.java:497) >> >>>>>>> >> >>>>>>> at >> >>>>>> >> >> org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317) >> >>>>>>> >> >>>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) >> >>>>>>> >> >>>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152) >> >>>>>>> >> >>>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> >>>> Method) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> >>>>>>> >> >>>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65) >> >>>>>>> >> >>>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org >> >>>>>>> >> >>>>>> >> >>>> >> >> >> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63) >> >>>>>>> >> >>>>>>> at >> org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714) >> >>>>>>> >> >>>>>>> at >> org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110) >> >>>>>>> >> >>>>>>> at scala.Option.map(Option.scala:145) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala) >> >>>>>>> >> >>>>>>> Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in >> >>>>>>> java.library.path >> >>>>>>> >> >>>>>>> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1864) >> >>>>>>> >> >>>>>>> at java.lang.Runtime.loadLibrary0(Runtime.java:870) >> >>>>>>> >> >>>>>>> at java.lang.System.loadLibrary(System.java:1122) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52) >> >>>>>>> >> >>>>>>> ... 49 more >> >>>>>>> >> >>>>>>> Exception in thread "main" >> >> java.lang.reflect.InvocationTargetException >> >>>>>>> >> >>>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> >>>> Method) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> >>>>>>> >> >>>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65) >> >>>>>>> >> >>>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org >> >>>>>>> >> >>>>>> >> >>>> >> >> >> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63) >> >>>>>>> >> >>>>>>> at >> org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714) >> >>>>>>> >> >>>>>>> at >> org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714) >> >>>>>>> >> >>>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110) >> >>>>>>> >> >>>>>>> at scala.Option.map(Option.scala:145) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala) >> >>>>>>> >> >>>>>>> Caused by: java.lang.IllegalArgumentException: >> >>>>>>> org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] >> null >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:171) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152) >> >>>>>>> >> >>>>>>> ... 38 more >> >>>>>>> >> >>>>>>> Caused by: org.xerial.snappy.SnappyError: >> >>>> [FAILED_TO_LOAD_NATIVE_LIBRARY] >> >>>>>>> null >> >>>>>>> >> >>>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229) >> >>>>>>> >> >>>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44) >> >>>>>>> >> >>>>>>> at >> >>>>>>> >> >>>>>> >> >>>> >> >> >> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169) >> >>>>>>> >> >>>>>>> ... 41 more >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO SparkContext: Invoking stop() from shutdown >> >> hook >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/metrics/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage/kill,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/api,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/static,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/executors/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/executors,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/environment/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/environment,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/storage/rdd/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/storage/rdd,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/storage/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/storage,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/pool/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/pool,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/stages,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/job/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/job,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/json,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped >> >>>>>>> o.s.j.s.ServletContextHandler{/jobs,null} >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO SparkUI: Stopped Spark web UI at >> >>>>>>> http://192.168.1.122:4040 >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO MapOutputTrackerMasterEndpoint: >> >>>>>>> MapOutputTrackerMasterEndpoint stopped! >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO MemoryStore: MemoryStore cleared >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO BlockManager: BlockManager stopped >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO BlockManagerMaster: BlockManagerMaster >> stopped >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO >> >>>>>>> OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: >> >>>>>>> OutputCommitCoordinator stopped! >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO SparkContext: Successfully stopped >> >> SparkContext >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Shutdown hook called >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO RemoteActorRefProvider$RemotingTerminator: >> >>>>>> Shutting >> >>>>>>> down remote daemon. >> >>>>>>> >> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory >> >>>>>>> >> >>>>>> >> >>>> >> >> >> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b >> >>>>>>> >> >>>>>>> -- >> >>>>>>> Thanks & Regards, >> >>>>>>> >> >>>>>>> *Rohit Jain* >> >>>>>>> Web developer | Consultant >> >>>>>>> Mob +91 8097283931 >> >>>>>> >> >>>>>> >> >>>>> >> >>>>> >> >>>>> -- >> >>>>> Thanks & Regards, >> >>>>> >> >>>>> *Rohit Jain* >> >>>>> Web developer | Consultant >> >>>>> Mob +91 8097283931 >> >>>> >> >>>> >> >>> >> >>> >> >>> -- >> >>> Thanks & Regards, >> >>> >> >>> *Rohit Jain* >> >>> Web developer | Consultant >> >>> Mob +91 8097283931 >> >> >> >> >> > >> > >> > -- >> > Thanks & Regards, >> > >> > *Rohit Jain* >> > Web developer | Consultant >> > Mob +91 8097283931 >> >> > > > -- > Thanks & Regards, > > *Rohit Jain* > Web developer | Consultant > Mob +91 8097283931 > -- Thanks & Regards, *Rohit Jain* Web developer | Consultant Mob +91 8097283931
