I've read through the mail archives and read the different threads.


I believe there is a great deal of value in teaching others.



I'm a 14yr vet of Java and would like to contribute to different Spark
projects. Here are my dilemmas:

1)    How does one quickly get a working environment up and running?



What do I mean by environment?


   1.  I have an IDE – Not a problem I can build using sbt.
   2. Environment to me is a working standalone spark cluster (Docker)
   which I can take what I build from #1 and re-deploy to test out my changes
   etc.
   3. What are the dependencies between projects internal to Spark?



How to on-board a new developer and make them productive as soon as
possible.



Not looking for answers to just these questions/dilemmas.



There is a wealth of knowledge here (existing Spark & sub-projects
developers/Architects).



My proposal, is to document the onboarding process and dependencies for a
new contributor, what someone will need in order to get working dev
environment up and running for the purposes of being able to add/test new
functionality to the Spark project. Also document how to setup a test
environment in order to deploy and test out said new functionality
(Docker/Standalone).





Suggestions?

Reply via email to