Starting a single Spark Slave (or Worker) -
when this
spark-1.3.0-bin-hadoop2.4% sbin/start-slave.sh
i message
failed launch org.apache.spark.deploy.worker.worker: default conf/spark-defaults.conf.
even though have this:
spark-1.3.0-bin-hadoop2.4% ll conf | grep spark-defaults.conf -rw-rwxr--+ 1 xxxx.xxxxx ama-unix 507 apr 29 07:09 spark-defaults.conf -rw-rwxr--+ 1 xxxx.xxxxx ama-unix 507 apr 13 12:06 spark-defaults.conf.template
any idea why?
thanks
first of all, should make sure using command correctly,
usage: start-slave.sh <worker#> <spark-master-url>
where <worker#>
worker number want launch on machine running script.
<spark-master-url>
spark://localhost:7077
Comments
Post a Comment