點此進入官網下載地址: 下載Spark
html
//下載好後用傳輸工具傳輸到咱們的linux上並解壓縮
tar -zxvf spark-2.2.0.tgz
根據Spark官方文檔可知,須要java8+,maven 3.3.9+版本java
若是不知道如何安裝,參考以下步驟:
(1)點此下載maven
(2)解壓縮:
tar -zxvf apache-maven-3.3.9-bin.tar.gz
(3)配置環境變量:
vim /etc/profile
export MAVEN_HOME=/usr/local/apache-maven-3.3.9
//根據機器配置爲maven分配儘量多的內存,可使得編譯更加快速
export MAVEN_OPTS="-Xmx2g -XX:ReservedCodeCacheSize=1024m"
export PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$PATH
//最後不要忘記source命令使得配置當即生效
source /etc/profile
linux
vim /etc/resolv.conf
namserver 8.8.8.8 nameserver 8.8.4.4
//檢驗是否成功連接外網
ping www.baidu.com
shell
cd /usr/local/spark-2.2.0
./dev/make-distribution.sh --name custom-spark --tgz -Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn
apache
編譯須要40分鐘左右:成功後在spark-2.2.0目錄下出現spark-2.2.0-bin-custom-spark.tgz包
將它移動到其餘目錄下:mv spark-2.2.0-bin-custom-spark.tgz /usr/local/
vim
tar -zxvf scala-2.11.8.tgz
vim /etc/profile
export SCALA_HOME=/usr/local/scala-2.11.8
export PATH=$SCALA_HOME/bin:$PATH
source /etc/profile
mv spark-env.sh.template spark-env.sh
mv slaves.template slaves
export JAVA_HOME=/usr/local/jdk1.8
export HADOOP_CONF_DIR=/usr/local/hadoop-2.6.0-cdh5.11.1/etc/hadoop
export SCALA_HOME=/usr/local/scala-2.11.8
master
slave1
slave2
scp -r spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1 slave1:/usr/local/
scp -r spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1 slave2:/usr/local/
cd /usr/local/spark-2.2.0-bin-hadoop2.6.0-cdh5.11.1/
bin/spark-shell
maven
出現以下界面便可:工具