Hadoop集羣搭建-04安裝配置HDFS

Hadoop集羣搭建-05安裝配置YARNhtml

Hadoop集羣搭建-04安裝配置HDFS node

Hadoop集羣搭建-03編譯安裝hadoopapache

Hadoop集羣搭建-02安裝配置Zookeeperbootstrap

Hadoop集羣搭建-01前期準備瀏覽器

HDFS是配合Hadoop使用的分佈式文件系統,分爲bash

namenode: nn1.hadoop nn2.hadoopssh

datanode: s1.hadoop s2.hadoop s3.hadoop分佈式

(看不明白這5臺虛擬機的請看前面 01前期準備 )oop

解壓配置文件ui

[hadoop@nn1 hadoop_base_op]$ ./ssh_all.sh mv /usr/local/hadoop/etc/hadoop /usr/local/hadoop/etc/hadoop_back

[hadoop@nn1 hadoop_base_op]$ ./scp_all.sh ../up/hadoop.tar.gz /tmp/

[hadoop@nn1 hadoop_base_op]$ #批量將自定義配置 壓縮包解壓到/usr/local/hadoop/etc/

#批量檢查配置是否正確解壓
[hadoop@nn1 hadoop_base_op]$ ./ssh_all.sh head /usr/local/hadoop/etc/hadoop/hadoop-env.sh
[hadoop@nn1 hadoop_base_op]$ ./ssh_root.sh chmown -R hadoop:hadoop /usr/local/hadoop/etc/hadoop
[hadoop@nn1 hadoop_base_op]$ ./ssh_root.sh chmod -R 770 /usr/local/hadoop/etc/hadoop

初始化HDFS

流程:

  1. 啓動zookeeper
  2. 啓動journalnode
  3. 啓動zookeeper客戶端,初始化HA的zookeeper信息
  4. 對nn1上的namenode進行格式化
  5. 啓動nn1上的namenode
  6. 在nn2上啓動同步namenode
  7. 啓動nn2上的namenode
  8. 啓動ZKFC
  9. 啓動dataname

 1.查看zookeeper狀態

[hadoop@nn1 zk_op]$ ./zk_ssh_all.sh /usr/local/zookeeper/bin/zkServer.sh status
ssh hadoop@"nn1.hadoop" "/usr/local/zookeeper/bin/zkServer.sh status"
ZooKeeper JMX enabled by default
Using config: /usr/local/zookeeper/bin/../conf/zoo.cfg
Mode: follower
OK!
ssh hadoop@"nn2.hadoop" "/usr/local/zookeeper/bin/zkServer.sh status"
ZooKeeper JMX enabled by default
Using config: /usr/local/zookeeper/bin/../conf/zoo.cfg
Mode: leader
OK!
ssh hadoop@"s1.hadoop" "/usr/local/zookeeper/bin/zkServer.sh status"
ZooKeeper JMX enabled by default
Using config: /usr/local/zookeeper/bin/../conf/zoo.cfg
Mode: follower
OK!

看到兩個follower和一個leader說明正常運行,若是沒有,就用下邊的命令啓動

[hadoop@nn1 zk_op]$ ./zk_ssh_all.sh /usr/local/zookeeper/bin/zkServer.sh start

2.啓動journalnode

這個玩意就是namenode的同步器。

#在nn1上啓動journalnode
[hadoop@nn1 zk_op]$ hadoop-daemon.sh start journalnode 

#在nn2上啓動journalnode
[hadoop@nn1 zk_op]$ hadoop-daemon.sh start journalnode

#能夠分別打開log來查看啓動狀態
[hadoop@nn1 zk_op]$ tail /usr/local/hadoop-2.7.3/logs/hadoop-hadoop-journalnode-nn1.hadoop.log
2019-07-22 17:15:54,164 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 8485
2019-07-22 17:15:54,190 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2019-07-22 17:15:54,191 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 8485: starting
#發現IPC通訊已經創建起來了,journalnode進程在8485

3.初始化HA信息(僅第一次運行,之後不須要)

[hadoop@nn1 zk_op]$ hdfs zkfc -formatZK
[hadoop@nn1 zk_op]$ /usr/local/zookeeper/bin/zkCli.sh
[zk: localhost:2181(CONNECTED) 0] ls /
[zookeeper, hadoop-ha]
[zk: localhost:2181(CONNECTED) 1] quit
Quitting...

4.對nn1上的namenode進行格式化(僅第一次運行,之後不須要)

[hadoop@nn1 zk_op]$ hadoop namenode -format
#出現下邊的說明初始化成功
#19/07/22 17:23:09 INFO common.Storage: Storage directory /data/dfsname has been successfully formatted.

5.啓動nn1的namenode

[hadoop@nn1 zk_op]$ hadoop-daemon.sh start namenode
[hadoop@nn1 zk_op]$ tail /usr/local/hadoop/logs/hadoop-hadoop-namenode-nn1.hadoop.log
#
#2019-07-22 17:24:57,321 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
#2019-07-22 17:24:57,322 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 9000: starting
#2019-07-22 17:24:57,385 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at: nn1.hadoop/192.168.10.6:9000
#2019-07-22 17:24:57,385 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services required for standby state
#2019-07-22 17:24:57,388 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Will roll logs on active node at nn2.hadoop/192.168.10.7:9000 every 120 seconds.
#2019-07-22 17:24:57,394 INFO org.apache.hadoop.hdfs.server.namenode.ha.StandbyCheckpointer: Starting standby checkpoint thread...
#Checkpointing active NN at http://nn2.hadoop:50070
#Serving checkpoints at http://nn1.hadoop:50070

6.在nn2機器上同步nn1的namenode狀態(僅第一次運行,之後不須要)

咱們來到nn2的控制檯!

###########必定要在nn2機器上運行這個!!!!############
[hadoop@nn2 ~]$ hadoop namenode -bootstrapStandby
=====================================================
About to bootstrap Standby ID nn2 from:
           Nameservice ID: ns1
        Other Namenode ID: nn1
  Other NN's HTTP address: http://nn1.hadoop:50070
  Other NN's IPC  address: nn1.hadoop/192.168.10.6:9000
             Namespace ID: 1728347664
            Block pool ID: BP-581543280-192.168.10.6-1563787389190
               Cluster ID: CID-42d2124d-9f54-4902-aa31-948fb0233943
           Layout version: -63
       isUpgradeFinalized: true
=====================================================
19/07/22 17:30:24 INFO common.Storage: Storage directory /data/dfsname has been successfully formatted.

7.啓動nn2的namenode

仍是在nn2控制檯運行!!

[hadoop@nn2 ~]$ hadoop-daemon.sh start namenode
#查看log來看看有沒有啓動成功
[hadoop@nn2 ~]$ tail /usr/local/hadoop-2.7.3/logs/hadoop-hadoop-namenode-nn2.hadoop.log

 8.啓動ZKFC

這時候在nn1和nn2分別啓動ZKFC,這時候兩臺機器的namenode,一個變成active一個變成standby!!ZKFC實現了HA高可用的自動切換!!

#############在nn1運行#################
[hadoop@nn1 zk_op]$ hadoop-daemon.sh start zkfc
#############在nn2運行####################
[hadoop@nn2 zk_op]$ hadoop-daemon.sh start zkfc

這時候在瀏覽器輸入地址訪問兩臺機器的hadoop界面

http://192.168.10.6:50070/dfshealth.html#tab-overview

http://192.168.10.7:50070/dfshealth.html#tab-overview

這兩個有一個active有一個是standby狀態。

9.啓動dataname就是啓動後三臺機器

########首先肯定slaves文件裏存放了須要配置誰爲datanode
[hadoop@nn1 hadoop]$ cat slaves
s1.hadoop
s2.hadoop
s3.hadoop
###########在顯示爲active的機器上運行##############
[hadoop@nn1 zk_op]$ hadoop-daemons.sh start datanode

10.查看硬盤容量

打開剛纔hadoop網頁,查看hdfs的硬盤格式化好了沒有。

這裏是HDFS系統爲每臺實體機器的硬盤默認預留了2G(能夠在配置文件hdfs-site.xml裏更改),而後實際用來作hdfs的是每臺機器15G,因此三臺一共45G。

如圖成功配置好HDFS。

 

以前寫的文章在這裏:

Hadoop集羣搭建-03編譯安裝hadoop

Hadoop集羣搭建-02安裝配置Zookeeper

Hadoop集羣搭建-01前期準備

相關文章
相關標籤/搜索