如下操做皆基於Ubuntu 16.04 32位環境 & Hadoop 2.7.3版本html
$ sudo apt update $ sudo apt install -y default-jdk $ sudo apt install -y vim $ sudo apt install -y openssh-server
/usr/lib/jvm/default-jdk
目錄/etc/profile
中添加下邊兩行內容並保存退出export JAVA_HOME=/usr/lib/jvm/default-java export PATH=$PATH:$JAVA_HOME/bin
$ sudo addgroup hadoop $ sudo adduser --ingroup hadoop hduser $ sudo usermod -a -G sudo hduser
$ ssh-keygen -t rsa -P "" $ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
$ ssh localhost
/home/hduser
$ tar -zxvf hadoop-2.7.3.tar.gz $ mv hadoop-2.7.3/ hadoop/ $ sudo chown -R hduser:hadoop hadoop
/etc/profile
,可是使用.bashrc更加靈活)export JAVA_HOME=/usr/lib/jvm/default-java export HADOOP_HOME=/home/hduser/hadoop export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin
~/hadoop/etc/hadoop/hadoop-env.sh
,在文件中添加下邊的內容export JAVA_HOME=/usr/lib/jvm/default-java
~/hadoop/etc/hadoop/yarn-env.sh
,在文件中添加下邊的內容export JAVA_HOME=/usr/lib/jvm/default-java
~/hadoop/etc/hadoop/core-site.xml
$ sudo mkdir -p /home/hduser/tmp $ sudo chown hduser:hadoop /home/hduser/tmp $ sudo chmod 750 /home/hduser/tmp
<configuration> <property> <name>fs.defaultFS</name> <value>hdfs://master:9000</value> </property> <property> <name>io.file.buffer.size</name> <value>131072</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/home/hduser/tmp</value> <description>Abase for other temporary directories.</description> </property> <property> <name>hadoop.proxyuser.hduser.hosts</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.hduser.groups</name> <value>*</value> </property> </configuration>
~/hadoop/etc/hadoop/hdfs-site.xml
,添加如下內容(各個屬性的內容能夠按照實際狀況修改)
$ sudo mkdir -p /home/hduser/dfs/name $ sudo chown hduser:hadoop /home/hduser/dfs/name $ sudo chmod 750 /home/hduser/dfs/name $ sudo mkdir -p /home/hduser/dfs/data $ sudo chown hduser:hadoop /home/hduser/dfs/data $ sudo chmod 750 /home/hduser/dfs/data
<configuration> <property> <name>dfs.namenode.secondary.http-address</name> <value>master:9001</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/home/hduser/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/home/hduser/dfs/data</value> </property> <property> <name>dfs.replication</name> <value>3</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property> </configuration>
~/hadoop/etc/hadoop/mapred-site.xml
,添加如下內容(各個屬性的內容能夠按照實際狀況修改)<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>master:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>master:19888</value> </property> </configuration>
~/hadoop/etc/hadoop/yarn-site.xml
(各個屬性的內容能夠按照實際狀況修改)<configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.address</name> <value> master:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value> master:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value> master:8031</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value> master:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value> master:8088</value> </property> </configuration>
~/hadoop/bin
目錄下,執行如下操做$ ./hdfs namenode –format
~/hadoop/sbin
執行如下命令$ ./start-dfs.sh $ ./start-yarn.sh
jps
命令能夠看到如下結果,說明啓動成功13058 Jps 13026 NodeManager 12916 ResourceManager 12169 DataNode
此處默認master和slave已經按照單機節點配置完成java
/etc/hosts
中寫入如下內容192.168.0.1 master 192.168.0.2 slave
$ ssh-copy-id -i ~/.ssh/id_rsa.pub hduser@slave
$ ssh-copy-id -i ~/.ssh/id_rsa.pub hduser@master
~/hadoop/etc/hadoop/slaves
master slave
~/hadoop/bin
目錄下,執行如下操做$ ./hdfs namenode –format
~/hadoop/sbin
執行如下命令$ ./start-dfs.sh $ ./start-yarn.sh
jps
命令能夠看到如下結果,說明啓動成功
4048 Jps 3899 ResourceManager 4013 NodeManager 3726 SecondaryNameNode 3423 NameNode 3535 DataNode
2754 DataNode 3012 Jps 2903 NodeManager
/home/hduser
$ tar -zxvf spark-2.3.1-bin-hadoop2.7.tgz
~/.bashrc
文件
export SPARK_HOME=/home/hduser/spark export PATH=$PATH:$SPARK_HOME/bin
~/spark/conf/slaves
文件
master slave
~/spark/sbin
中執行如下命令./start-all.sh
jps
,發現如下結果說明spark運行成功5257 Worker 5177 Master 3726 SecondaryNameNode 5326 Jps 3432 NameNode 3535 DataNode
jsp
,發現一下結果說明spark運行成功2754 DataNode 3717 Worker 3771 Jps
/home/hduser
$ tar -zxvf scala-2.12.6.tgz
~/.bashrc
文件
export SCALA_HOME=/home/hduser/scala export PATH=$PATH:$SCALA_HOME/bin
$ sudo apt install eclipse-platform