点此进入官网下载地址: 下载Spark
html
//下载好后用传输工具传输到咱们的linux上并解压缩
tar -zxvf spark-2.2.0.tgz
根据Spark官方文档可知,须要java8+,maven 3.3.9+版本java
若是不知道如何安装,参考以下步骤:
(1)点此下载maven
(2)解压缩:
tar -zxvf apache-maven-3.3.9-bin.tar.gz
(3)配置环境变量:
vim /etc/profile
export MAVEN_HOME=/usr/local/apache-maven-3.3.9
//根据机器配置为maven分配尽量多的内存,可使得编译更加快速
export MAVEN_OPTS="-Xmx2g -XX:ReservedCodeCacheSize=1024m"
export PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$PATH
//最后不要忘记source命令使得配置当即生效
source /etc/profile
linux
vim /etc/resolv.conf
namserver 8.8.8.8 nameserver 8.8.4.4
//检验是否成功连接外网
ping www.baidu.com
shell
cd /usr/local/spark-2.2.0
./dev/make-distribution.sh --name custom-spark --tgz -Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn
apache
编译须要40分钟左右:成功后在spark-2.2.0目录下出现spark-2.2.0-bin-custom-spark.tgz包
将它移动到其余目录下:mv spark-2.2.0-bin-custom-spark.tgz /usr/local/
vim
tar -zxvf scala-2.11.8.tgz
vim /etc/profile
export SCALA_HOME=/usr/local/scala-2.11.8
export PATH=$SCALA_HOME/bin:$PATH
source /etc/profile
mv spark-env.sh.template spark-env.sh
mv slaves.template slaves
export JAVA_HOME=/usr/local/jdk1.8
export HADOOP_CONF_DIR=/usr/local/hadoop-2.6.0-cdh5.11.1/etc/hadoop
export SCALA_HOME=/usr/local/scala-2.11.8
master
slave1
slave2
scp -r spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1 slave1:/usr/local/
scp -r spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1 slave2:/usr/local/
cd /usr/local/spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1/
bin/spark-shell
maven
出现以下界面便可:工具