Tried pyspark command on same machine which uses Anaconda python and sc.version returned value.
Zeppelin: zeppelin.python /home/cloudera/anaconda2/bin In zeppelin, nothing is returned. On Wed, Sep 14, 2016 at 11:53 AM, moon soo Lee <m...@apache.org> wrote: > Did you export SPARK_HOME in conf/zeppelin-env.sh? > Could you verify the some code works with ${SPARK_HOME}/bin/pyspark, on > the same machine that zeppelin runs? > > Thanks, > moon > > > On Wed, Sep 14, 2016 at 8:07 AM Abhi Basu <9000r...@gmail.com> wrote: > >> Oops sorry. the above code generated this error: >> >> RROR [2016-09-14 10:04:27,121] ({qtp2003293121-11} >> NotebookServer.java[onMessage]:221) - Can't handle message >> org.apache.zeppelin.interpreter.InterpreterException: >> org.apache.thrift.transport.TTransportException >> at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel( >> RemoteInterpreter.java:319) >> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.cancel( >> LazyOpenInterpreter.java:100) >> at org.apache.zeppelin.notebook.Paragraph.jobAbort(Paragraph.java:330) >> at org.apache.zeppelin.scheduler.Job.abort(Job.java:239) >> at org.apache.zeppelin.socket.NotebookServer.cancelParagraph( >> NotebookServer.java:995) >> at org.apache.zeppelin.socket.NotebookServer.onMessage( >> NotebookServer.java:180) >> at org.apache.zeppelin.socket.NotebookSocket.onWebSocketText( >> NotebookSocket.java:56) >> at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver. >> onTextMessage(JettyListenerEventDriver.java:128) >> at org.eclipse.jetty.websocket.common.message.SimpleTextMessage. >> messageComplete(SimpleTextMessage.java:69) >> at org.eclipse.jetty.websocket.common.events.AbstractEventDriver. >> appendMessage(AbstractEventDriver.java:65) >> at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver. >> onTextFrame(JettyListenerEventDriver.java:122) >> at org.eclipse.jetty.websocket.common.events.AbstractEventDriver. >> incomingFrame(AbstractEventDriver.java:161) >> at org.eclipse.jetty.websocket.common.WebSocketSession.incomingFrame( >> WebSocketSession.java:309) >> at org.eclipse.jetty.websocket.common.extensions. >> ExtensionStack.incomingFrame(ExtensionStack.java:214) >> at org.eclipse.jetty.websocket.common.Parser.notifyFrame(Parser.java:220) >> at org.eclipse.jetty.websocket.common.Parser.parse(Parser.java:258) >> at org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection. >> readParse(AbstractWebSocketConnection.java:632) >> at org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection. >> onFillable(AbstractWebSocketConnection.java:480) >> at org.eclipse.jetty.io.AbstractConnection$2.run( >> AbstractConnection.java:544) >> at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob( >> QueuedThreadPool.java:635) >> at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run( >> QueuedThreadPool.java:555) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: org.apache.thrift.transport.TTransportException >> at org.apache.thrift.transport.TIOStreamTransport.read( >> TIOStreamTransport.java:132) >> at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) >> at org.apache.thrift.protocol.TBinaryProtocol.readAll( >> TBinaryProtocol.java:429) >> at org.apache.thrift.protocol.TBinaryProtocol.readI32( >> TBinaryProtocol.java:318) >> at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin( >> TBinaryProtocol.java:219) >> at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69) >> at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$ >> Client.recv_cancel(RemoteInterpreterService.java:274) >> at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$ >> Client.cancel(RemoteInterpreterService.java:259) >> at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel( >> RemoteInterpreter.java:316) >> ... 21 more >> >> >> This is my spark interpreter settings: >> >> >> spark %spark , %spark.pyspark , %spark.r , %spark.sql , %spark.dep >> Option >> Interpreter for note >> >> Connect to existing process >> Properties >> name value >> args >> master yarn-client >> spark.app.name Zeppelin >> spark.cores.max >> spark.executor.memory >> zeppelin.R.cmd R >> zeppelin.R.image.width 100% >> zeppelin.R.knitr true >> zeppelin.R.render.options out.format = 'html', comment = NA, echo = >> FALSE, results = 'asis', message = F, warning = F >> zeppelin.dep.additionalRemoteRepository spark-packages,http://dl. >> bintray.com/spark-packages/maven,false; >> zeppelin.dep.localrepo local-repo >> zeppelin.interpreter.localRepo /usr/local/bin/zeppelin-0.6.1- >> bin-all/local-repo/2BXF675WU >> zeppelin.pyspark.python python >> zeppelin.spark.concurrentSQL false >> zeppelin.spark.importImplicit true >> zeppelin.spark.maxResult 1000 >> zeppelin.spark.printREPLOutput true >> zeppelin.spark.sql.stacktrace false >> zeppelin.spark.useHiveContext true >> >> >> On Wed, Sep 14, 2016 at 10:05 AM, Abhi Basu <9000r...@gmail.com> wrote: >> >>> %pyspark >>> >>> input_file = "hdfs:////tmp/filenname.gz" >>> >>> raw_rdd = sc.textFile(input_file) >>> >>> >>> >> >> >> -- >> Abhi Basu >> > -- Abhi Basu