[Spark] Spark 安裝配置

原文地址:http://www.javashuo.com/article/p-anksadci-ew.htmlhtml

Windows 上的單機版安裝

下載地址:http://spark.apache.org/downloads.htmljava

本文以 Spark 2.3.0 爲例shell

a1dbaff8-6f81-446a-b62c-d2f38c52bb5c

>>> 下載下來的文件是 tgz 格式的壓縮文件,直接利用壓縮軟件將其打開,能夠看見裏面有一個 tar 格式的壓縮文件,繼續用壓縮軟件打開,最終以下圖所示:express

399ae4e4-2528-4769-9621-7f00825d14ec

>>> 將其中的 spark-2.3.0-bin-hadoop2.7 文件夾解壓,新增 SPARK_HOME 環境變量,設置爲解壓路徑,並將其追加到 PATH 環境變量;apache

e0b1ea30-ee34-4c1d-af17-78d86664d066

d4010648-fc49-4d04-a660-c6010709b387

>>> 此時在 cmd 窗口中執行 "spark-shell" 命令可獲得以下提示:bash

C:\Users\Memento>spark-shell
2018-06-06 23:39:36 ERROR Shell:397 - Failed to locate the winutils binary in the hadoop binary path
java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries.
        at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:379)
        at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:394)
        at org.apache.hadoop.util.Shell.<clinit>(Shell.java:387)
        at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:80)
        at org.apache.hadoop.security.SecurityUtil.getAuthenticationMethod(SecurityUtil.java:611)
        at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:273)
        at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:261)
        at org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:791)
        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:761)
        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:634)
        at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2464)
        at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2464)
        at scala.Option.getOrElse(Option.scala:121)
        at org.apache.spark.util.Utils$.getCurrentUserName(Utils.scala:2464)
        at org.apache.spark.SecurityManager.<init>(SecurityManager.scala:222)
        at org.apache.spark.deploy.SparkSubmit$.secMgr$lzycompute$1(SparkSubmit.scala:393)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$secMgr$1(SparkSubmit.scala:393)
        at org.apache.spark.deploy.SparkSubmit$$anonfun$prepareSubmitEnvironment$7.apply(SparkSubmit.scala:401)
        at org.apache.spark.deploy.SparkSubmit$$anonfun$prepareSubmitEnvironment$7.apply(SparkSubmit.scala:401)
        at scala.Option.map(Option.scala:146)
        at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:400)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:170)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:136)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
2018-06-06 23:39:36 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Spark context Web UI available at http://Memento-PC:4040
Spark context available as 'sc' (master = local[*], app id = local-1528299586814).
Spark session available as 'spark'.
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 2.3.0
      /_/

Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_151)
Type in expressions to have them evaluated.
Type :help for more information.

提示說在 hadoop 路徑下沒法定位到 winutils,因此接下來須要配置 Hadoop;session

>>> 詳見:Windows 下的 Hadoop 2.7.5 環境搭建app

>>> 隨後再從新執行 "spark-shell" 命令便可;oop

06e5a299-f366-4474-8367-ae45837c3941

 

By. Mementoui

相關文章
相關標籤/搜索