計算圓周率
# bin/spark-submit \ --class org.apache.spark.examples.SparkPi \ --executor-memory 1G \ --total-executor-cores 2 \ ./examples/jars/spark-examples_2.11-2.3.3.jar \ 100
wordcountjava
sc.textFile("input").flatMap(_.split(" ")).map((_,1)).reduceByKey(_+_).collect
輸出python
res3: Array[(String, Int)] = Array((scala,2), ("",1), (object,1), (is,1), (python,3), (hello,4), (java,3), (go,1), (and,1))
spark 運行在yarnshell
mv spark-env.sh.template spark-env.sh # 編輯 vim spark-env.sh #修改 YARN_CONF_DIR=/itcatst/hadoop2.8/etc/hadoop #執行命令 測試 bin/spark-submit \ --class org.apache.spark.examples.SparkPi \ --master yarn \ --deploy-mode client \ ./examples/jars/spark-examples_2.11-2.3.3.jar \ 100
啓動apache
bin/spark-shell --master yarn
local模式vim
bin/spark-submit \ --class org.wondersgroup.bigdata.spark.WordCount \ WordCount-jar-with-dependencies.jar
獨立的部署模式oop
mv slaves.template slaves #編輯 vim slaves hadoop01 hadoop02 hadoop03 vim spark-env.sh export JAVA_HOME=/usr/local/jdk1.8.0_221 SPARK_MASTER_HOST=hadoop01 SPARK_MASTER_PORT=7077
計算測試
bin/spark-submit \ --class org.apache.spark.examples.SparkPi \ --master spark://hadoop01:7077 \ --executor-memory 1G \ --total-executor-cores 2 \ ./examples/jars/spark-examples_2.11-2.3.3.jar \ 100
啓動spark集羣spa
sbin/spark-all.sh