Thanks.
I've set SPARK_HOME and SPARK_CONF_DIR appropriately in .bash_profile
But when I start worker like this
spark-1.3.0-bin-hadoop2.4/sbin/start-slave.sh
I still get
failed to launch org.apache.spark.deploy.worker.Worker:
Default is conf/spark-defaults.conf.
15/04/27 11:51:33 DEBUG Utils: Shutdown hook called
On Mon, Apr 27, 2015 at 1:15 PM, Zoltán Zvara <[email protected]>
wrote:
> You should distribute your configuration file to workers and set the
> appropriate environment variables, like HADOOP_HOME, SPARK_HOME,
> HADOOP_CONF_DIR, SPARK_CONF_DIR.
>
> On Mon, Apr 27, 2015 at 12:56 PM James King <[email protected]> wrote:
>
>> I renamed spark-defaults.conf.template to spark-defaults.conf
>> and invoked
>>
>> spark-1.3.0-bin-hadoop2.4/sbin/start-slave.sh
>>
>> But I still get
>>
>> failed to launch org.apache.spark.deploy.worker.Worker:
>> --properties-file FILE Path to a custom Spark properties file.
>> Default is conf/spark-defaults.conf.
>>
>> But I'm thinking it should pick up the default spark-defaults.conf from
>> conf dir
>>
>> Am I expecting or doing something wrong?
>>
>> Regards
>> jk
>>
>>
>>