Tried pyspark command on same machine which uses Anaconda python and
sc.version returned value.

Zeppelin:
zeppelin.python /home/cloudera/anaconda2/bin

In zeppelin, nothing is returned.


On Wed, Sep 14, 2016 at 11:53 AM, moon soo Lee <m...@apache.org> wrote:

> Did you export SPARK_HOME in conf/zeppelin-env.sh?
> Could you verify the some code works with ${SPARK_HOME}/bin/pyspark, on
> the same machine that zeppelin runs?
>
> Thanks,
> moon
>
>
> On Wed, Sep 14, 2016 at 8:07 AM Abhi Basu <9000r...@gmail.com> wrote:
>
>> Oops sorry. the above code generated this error:
>>
>> RROR [2016-09-14 10:04:27,121] ({qtp2003293121-11}
>> NotebookServer.java[onMessage]:221) - Can't handle message
>> org.apache.zeppelin.interpreter.InterpreterException:
>> org.apache.thrift.transport.TTransportException
>> at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel(
>> RemoteInterpreter.java:319)
>> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.cancel(
>> LazyOpenInterpreter.java:100)
>> at org.apache.zeppelin.notebook.Paragraph.jobAbort(Paragraph.java:330)
>> at org.apache.zeppelin.scheduler.Job.abort(Job.java:239)
>> at org.apache.zeppelin.socket.NotebookServer.cancelParagraph(
>> NotebookServer.java:995)
>> at org.apache.zeppelin.socket.NotebookServer.onMessage(
>> NotebookServer.java:180)
>> at org.apache.zeppelin.socket.NotebookSocket.onWebSocketText(
>> NotebookSocket.java:56)
>> at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.
>> onTextMessage(JettyListenerEventDriver.java:128)
>> at org.eclipse.jetty.websocket.common.message.SimpleTextMessage.
>> messageComplete(SimpleTextMessage.java:69)
>> at org.eclipse.jetty.websocket.common.events.AbstractEventDriver.
>> appendMessage(AbstractEventDriver.java:65)
>> at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.
>> onTextFrame(JettyListenerEventDriver.java:122)
>> at org.eclipse.jetty.websocket.common.events.AbstractEventDriver.
>> incomingFrame(AbstractEventDriver.java:161)
>> at org.eclipse.jetty.websocket.common.WebSocketSession.incomingFrame(
>> WebSocketSession.java:309)
>> at org.eclipse.jetty.websocket.common.extensions.
>> ExtensionStack.incomingFrame(ExtensionStack.java:214)
>> at org.eclipse.jetty.websocket.common.Parser.notifyFrame(Parser.java:220)
>> at org.eclipse.jetty.websocket.common.Parser.parse(Parser.java:258)
>> at org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection.
>> readParse(AbstractWebSocketConnection.java:632)
>> at org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection.
>> onFillable(AbstractWebSocketConnection.java:480)
>> at org.eclipse.jetty.io.AbstractConnection$2.run(
>> AbstractConnection.java:544)
>> at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(
>> QueuedThreadPool.java:635)
>> at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(
>> QueuedThreadPool.java:555)
>> at java.lang.Thread.run(Thread.java:745)
>> Caused by: org.apache.thrift.transport.TTransportException
>> at org.apache.thrift.transport.TIOStreamTransport.read(
>> TIOStreamTransport.java:132)
>> at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
>> at org.apache.thrift.protocol.TBinaryProtocol.readAll(
>> TBinaryProtocol.java:429)
>> at org.apache.thrift.protocol.TBinaryProtocol.readI32(
>> TBinaryProtocol.java:318)
>> at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(
>> TBinaryProtocol.java:219)
>> at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
>> at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$
>> Client.recv_cancel(RemoteInterpreterService.java:274)
>> at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$
>> Client.cancel(RemoteInterpreterService.java:259)
>> at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel(
>> RemoteInterpreter.java:316)
>> ... 21 more
>>
>>
>> This is my spark interpreter settings:
>>
>>
>> spark %spark , %spark.pyspark , %spark.r , %spark.sql , %spark.dep
>> Option
>> Interpreter for note
>>
>> Connect to existing process
>> Properties
>> name value
>> args
>> master yarn-client
>> spark.app.name Zeppelin
>> spark.cores.max
>> spark.executor.memory
>> zeppelin.R.cmd R
>> zeppelin.R.image.width 100%
>> zeppelin.R.knitr true
>> zeppelin.R.render.options out.format = 'html', comment = NA, echo =
>> FALSE, results = 'asis', message = F, warning = F
>> zeppelin.dep.additionalRemoteRepository spark-packages,http://dl.
>> bintray.com/spark-packages/maven,false;
>> zeppelin.dep.localrepo local-repo
>> zeppelin.interpreter.localRepo /usr/local/bin/zeppelin-0.6.1-
>> bin-all/local-repo/2BXF675WU
>> zeppelin.pyspark.python python
>> zeppelin.spark.concurrentSQL false
>> zeppelin.spark.importImplicit true
>> zeppelin.spark.maxResult 1000
>> zeppelin.spark.printREPLOutput true
>> zeppelin.spark.sql.stacktrace false
>> zeppelin.spark.useHiveContext true
>>
>>
>> On Wed, Sep 14, 2016 at 10:05 AM, Abhi Basu <9000r...@gmail.com> wrote:
>>
>>> %pyspark
>>>
>>> input_file = "hdfs:////tmp/filenname.gz"
>>>
>>> raw_rdd = sc.textFile(input_file)
>>>
>>>
>>>
>>
>>
>> --
>> Abhi Basu
>>
>


-- 
Abhi Basu

Reply via email to