I am unable to run hive scripts in Spark 1.1.0 pro-grammatically in hadoop prompt but I could do it manually. Can anyone help me to run hive scripts pro-grammatically in spark1.1.0 cluster on EMR?
Manual running steps:- hadoop@ip-10-151-71-224:~/tmpSpark/spark1.1/spark$ ./bin/spark-shell --driver-memory 4G --executor-memory 4G Spark assembly has been built with Hive, including Datanucleus jars on classpath Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 14/09/26 11:54:29 INFO SecurityManager: Changing view acls to: hadoop, 14/09/26 11:54:29 INFO SecurityManager: Changing modify acls to: hadoop, 14/09/26 11:54:29 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop, ); users with modify permissions: Set(hadoop, ) 14/09/26 11:54:29 INFO HttpServer: Starting HTTP Server 14/09/26 11:54:29 INFO Utils: Successfully started service 'HTTP class server' on port 52081. Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_40) Type in expressions to have them evaluated. Type :help for more information. 14/09/26 11:54:34 INFO SecurityManager: Changing view acls to: hadoop, 14/09/26 11:54:34 INFO SecurityManager: Changing modify acls to: hadoop, 14/09/26 11:54:34 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop, ); users with modify permissions: Set(hadoop, ) 14/09/26 11:54:35 INFO Slf4jLogger: Slf4jLogger started 14/09/26 11:54:35 INFO Remoting: Starting remoting 14/09/26 11:54:35 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@ip-10-151-71-224.ec2.internal:46137] 14/09/26 11:54:35 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@ip-10-151-71-224.ec2.internal:46137] 14/09/26 11:54:35 INFO Utils: Successfully started service 'sparkDriver' on port 46137. 14/09/26 11:54:35 INFO SparkEnv: Registering MapOutputTracker 14/09/26 11:54:35 INFO SparkEnv: Registering BlockManagerMaster 14/09/26 11:54:35 INFO DiskBlockManager: Created local directory at /tmp/spark-local-20140926115435-fa1a 14/09/26 11:54:35 INFO Utils: Successfully started service 'Connection manager for block manager' on port 47623. 14/09/26 11:54:35 INFO ConnectionManager: Bound socket to port 47623 with id = ConnectionManagerId(ip-10-151-71-224.ec2.internal,47623) 14/09/26 11:54:35 INFO MemoryStore: MemoryStore started with capacity 2.1 GB 14/09/26 11:54:35 INFO BlockManagerMaster: Trying to register BlockManager 14/09/26 11:54:35 INFO BlockManagerMasterActor: Registering block manager ip-10-151-71-224.ec2.internal:47623 with 2.1 GB RAM 14/09/26 11:54:35 INFO BlockManagerMaster: Registered BlockManager 14/09/26 11:54:35 INFO HttpFileServer: HTTP File server directory is /tmp/spark-dc2260ea-18cc-4204-8f02-36bcc1df1126 14/09/26 11:54:35 INFO HttpServer: Starting HTTP Server 14/09/26 11:54:36 INFO Utils: Successfully started service 'HTTP file server' on port 49299. 14/09/26 11:54:41 INFO Utils: Successfully started service 'SparkUI' on port 4040. 14/09/26 11:54:41 INFO SparkUI: Started SparkUI at http://ip-10-151-71-224.ec2.internal:4040 14/09/26 11:54:41 INFO Executor: Using REPL class URI: http://10.151.71.224:52081 14/09/26 11:54:41 INFO AkkaUtils: Connecting to HeartbeatReceiver: akka.tcp://sparkDriver@ip-10-151-71-224.ec2.internal:46137/user/HeartbeatReceiver 14/09/26 11:54:41 INFO SparkILoop: Created spark context.. Spark context available as sc. scala> scala> val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc); hiveContext: org.apache.spark.sql.hive.HiveContext = org.apache.spark.sql.hive.HiveContext@3e77175c scala> hiveContext.hql("CREATE EXTERNAL TABLE IF NOT EXISTS test (time string, id string) ROW FORMAT DELIMITED STORED AS TEXTFILE LOCATION 's3n://output/test/'"); -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-run-hive-scripts-pro-grammatically-in-Spark-1-1-0-tp15225.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org