spark

計算圓周率
# bin/spark-submit \
--class org.apache.spark.examples.SparkPi \
--executor-memory 1G \
--total-executor-cores 2 \
./examples/jars/spark-examples_2.11-2.3.3.jar \
100  

wordcountjava

sc.textFile("input").flatMap(_.split(" ")).map((_,1)).reduceByKey(_+_).collect  

輸出python

res3: Array[(String, Int)] = Array((scala,2), ("",1), (object,1), (is,1), (python,3), (hello,4), (java,3), (go,1), (and,1))  

 spark 運行在yarnshell

mv spark-env.sh.template spark-env.sh
# 編輯
vim spark-env.sh 
#修改
YARN_CONF_DIR=/itcatst/hadoop2.8/etc/hadoop

#執行命令 測試
 bin/spark-submit \
--class org.apache.spark.examples.SparkPi \
--master yarn \
--deploy-mode client \
./examples/jars/spark-examples_2.11-2.3.3.jar \
100    

 啓動apache

bin/spark-shell --master yarn  

local模式vim

bin/spark-submit \
--class org.wondersgroup.bigdata.spark.WordCount \
WordCount-jar-with-dependencies.jar 

獨立的部署模式oop

mv slaves.template slaves
#編輯
vim slaves

hadoop01
hadoop02
hadoop03

vim spark-env.sh
export JAVA_HOME=/usr/local/jdk1.8.0_221
SPARK_MASTER_HOST=hadoop01
SPARK_MASTER_PORT=7077  

 計算測試

bin/spark-submit \ 
--class org.apache.spark.examples.SparkPi \
--master spark://hadoop01:7077 \
--executor-memory 1G \
--total-executor-cores 2 \
./examples/jars/spark-examples_2.11-2.3.3.jar \
100  

 啓動spark集羣spa

sbin/spark-all.sh  
相關文章
相關標籤/搜索