Hi,

I want to use pyspark as distributed via conda in headless mode.
It looks like the hadoop binaries are bundles (= pip distributes a default
version)
https://stackoverflow.com/questions/63661404/bootstrap-spark-itself-on-yarn.

I want to ask if it would be possible to A) distribute the headless version
(=without hadoop) instead or B) distribute the headless version
additionally for pip & conda-forge distribution channels.

Best,
Georg

Reply via email to