Spark Standalone Mode Configuration
For currently popular distributed framework Spark, here it shows the intro and steps to configure the spark standalone mode on several machines.
It is easy to configure it from stratch. The following instruction I note down is based on the spark-2.0.2-bin-hadoop2.7 as example on the linux debian machines for scala programming.
Assume you have two machines with IP: 192.168.0.51 and 192.168.0.52
1. Preinstall java, scala, sbt
check: https://www.scala-lang.org/download/install.html
http://www.scala-sbt.org/0.13/docs/Installing-sbt-on-Linux.html
2. Download prebuilt spark version with hadoop. or you can compile on your own
the link can be referenced: https://spark.apache.org/downloads.html
3. Unzip the file and create the link for easy visit later
$ ssh-keygen -t rsa -P ""
(2) Copy id_rsa.pub to authorized-keys
$ cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
(3) Start ssh localhost if you want to work in only one localhost machine for spark standalone
$ ssh localhost