Hi Anil, I generally create an anaconda environment, and then install pyspark in it, and then configure the interpreter to point to that particular environment. Never faced an issue with my approach.
Regards, Gourav Sengupta On Wed, Nov 17, 2021 at 7:39 AM Anil Kulkarni <anil...@gmail.com> wrote: > Hi Spark community, > > I am having a hard time setting up my Pycharm to work with pyspark. Can > any of you point me to documentation available? > > Things I have tried till now : > > 1. Download and Install Apache spark > 2. Add pyspark package in pycharm. > 3. Add SPARK_HOME. PYTHONPATH, HADOOP_HOME env variables to Run config > > Error I am getting : > > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > Setting default log level to "WARN". > To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use > setLogLevel(newLevel). > 21/11/16 23:26:28 WARN NativeCodeLoader: *Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable* > Traceback (most recent call last): > > > -- > Cheers, > Anil Kulkarni > https://anilkulkarni.com/ > >