You need to set proper permission for the directory: :/user/spark/applicationHistory
in your local file system Thanks Best Regards On Fri, Mar 27, 2015 at 2:18 AM, pzilaro <payalzil...@yahoo.com> wrote: > I get the following error message when I start pyspark shell. > The config has the following settings- > # spark.master spark://master:7077 > # spark.eventLog.enabled true > # spark.eventLog.dir hdfs://namenode:8021/directory > # spark.serializer org.apache.spark.serializer.KryoSerializer > spark.eventLog.dir=/user/spark/applicationHistory > spark.eventLog.enabled=true > spark.yarn.historyServer.address=name101-car.ldcint.com:10020 > > > [pzilaro@name101-car conf]$ pyspark > Python 2.6.6 (r266:84292, Jan 22 2014, 09:42:36) > [GCC 4.4.7 20120313 (Red Hat 4.4.7-4)] on linux2 > Type "help", "copyright", "credits" or "license" for more information. > 15/03/26 13:46:06 INFO spark.SecurityManager: Changing view acls to: > pzilaro > 15/03/26 13:46:06 INFO spark.SecurityManager: Changing modify acls to: > pzilaro > 15/03/26 13:46:06 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(pzilaro); users with modify permissions: Set(pzilaro) > 15/03/26 13:46:07 INFO slf4j.Slf4jLogger: Slf4jLogger started > 15/03/26 13:46:07 INFO Remoting: Starting remoting > 15/03/26 13:46:07 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkdri...@name101-car.ldcint.com:48040] > 15/03/26 13:46:07 INFO Remoting: Remoting now listens on addresses: > [akka.tcp://sparkdri...@name101-car.ldcint.com:48040] > 15/03/26 13:46:07 INFO util.Utils: Successfully started service > 'sparkDriver' on port 48040. > 15/03/26 13:46:07 INFO spark.SparkEnv: Registering MapOutputTracker > 15/03/26 13:46:07 INFO spark.SparkEnv: Registering BlockManagerMaster > 15/03/26 13:46:07 INFO storage.DiskBlockManager: Created local directory at > /tmp/spark-local-20150326134607-072e > 15/03/26 13:46:07 INFO storage.MemoryStore: MemoryStore started with > capacity 265.4 MB > 15/03/26 13:46:08 INFO spark.HttpFileServer: HTTP File server directory is > /tmp/spark-2f342a3a-c5bb-474d-867b-8bd5b9f9d1ac > 15/03/26 13:46:08 INFO spark.HttpServer: Starting HTTP Server > 15/03/26 13:46:08 INFO server.Server: jetty-8.y.z-SNAPSHOT > 15/03/26 13:46:08 INFO server.AbstractConnector: Started > SocketConnector@0.0.0.0:55296 > 15/03/26 13:46:08 INFO util.Utils: Successfully started service 'HTTP file > server' on port 55296. > 15/03/26 13:46:08 INFO server.Server: jetty-8.y.z-SNAPSHOT > 15/03/26 13:46:08 INFO server.AbstractConnector: Started > SelectChannelConnector@0.0.0.0:4040 > 15/03/26 13:46:08 INFO util.Utils: Successfully started service 'SparkUI' > on > port 4040. > 15/03/26 13:46:08 INFO ui.SparkUI: Started SparkUI at > http://name101-car.ldcint.com:4040 > 15/03/26 13:46:08 INFO util.AkkaUtils: Connecting to HeartbeatReceiver: > akka.tcp://sparkdri...@name101-car.ldcint.com:48040/user/HeartbeatReceiver > 15/03/26 13:46:08 INFO netty.NettyBlockTransferService: Server created on > 55241 > 15/03/26 13:46:08 INFO storage.BlockManagerMaster: Trying to register > BlockManager > 15/03/26 13:46:08 INFO storage.BlockManagerMasterActor: Registering block > manager localhost:55241 with 265.4 MB RAM, BlockManagerId(<driver>, > localhost, 55241) > 15/03/26 13:46:08 INFO storage.BlockManagerMaster: Registered BlockManager > Traceback (most recent call last): > File "/usr/lib/spark/python/pyspark/shell.py", line 45, in <module> > sc = SparkContext(appName="PySparkShell", pyFiles=add_files) > File "/usr/lib/spark/python/pyspark/context.py", line 105, in __init__ > conf, jsc) > File "/usr/lib/spark/python/pyspark/context.py", line 153, in _do_init > self._jsc = jsc or self._initialize_context(self._conf._jconf) > File "/usr/lib/spark/python/pyspark/context.py", line 201, in > _initialize_context > return self._jvm.JavaSparkContext(jconf) > File > "/usr/lib/spark/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py", line > 701, in __call__ > File "/usr/lib/spark/python/lib/py4j-0.8.2.1-src.zip/py4j/protocol.py", > line 300, in get_return_value > py4j.protocol.Py4JJavaError: An error occurred while calling > None.org.apache.spark.api.java.JavaSparkContext. > : java.io.IOException: Error in creating log directory: > file:/user/spark/applicationHistory//local-1427402768636 > at > org.apache.spark.util.FileLogger.createLogDir(FileLogger.scala:133) > at org.apache.spark.util.FileLogger.start(FileLogger.scala:115) > at > > org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:74) > at org.apache.spark.SparkContext.<init>(SparkContext.scala:353) > at > > org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:61) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) > at > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:234) > at > py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379) > at py4j.Gateway.invoke(Gateway.java:214) > at > > py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:79) > at > py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:68) > at py4j.GatewayConnection.run(GatewayConnection.java:207) > at java.lang.Thread.run(Thread.java:745) > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Error-in-creating-log-directory-tp22250.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >