1、下载scala-2.11.1,并解压到/usr/scala/scala-2.11.1
2、下载spark-2.0.0-bin-hadoop2.4,并解压到/home/hadoop/Deploy/spark-2.0.0
(*如果要看后续文章,建议使用hadoop-2.5.2 hbase-1.1.2 hive-1.2.1 spark-2.0.0)
3、复制spark-env.sh.template为spark-env.sh,并添加下面几行
export JAVA_HOME=/usr/java/jdk1.7.0_79 export SCALA_HOME=/usr/scala/scala-2.11.1/ export SPARK_MASTER_IP=hiup01 export SPARK_WORKER_MEMORY=1g export HADOOP_CONF_DIR=/home/hadoop/Deploy/hadoop-2.5.2/etc/hadoop
4、复制slaves.template为slaves,并添加下面几行
hiup01 hiup02 hiup03
5、将scala-2.11.1及spark-2.0.0复制到hiup02及hiup03
6、环境搭建完毕。