hadoop 徹底分佈式部署

一.實驗環境準備
  須要準備四臺Linux操做系統的服務器,配置參數最好同樣,因爲個人虛擬機是以前僞分佈式部署而來的,所以個人環境都一致,而且天天虛擬機默認都是Hadoop僞分佈式喲!
1>.NameNode服務器(172.20.20.228)node

2>.DataNode服務器(172.20.20.226-220)shell

  二.修改Hadoop的配置文件服務器

  修改的配置文件路徑是我以前拷貝的full目錄,絕對路徑是:「/tosp/opt/hadoop」,修改這個目錄下的文件以後,咱們將hadoop目錄鏈接過來便可,當你須要僞分佈式或者本地模式的時候只須要改變軟鏈接指向的目錄便可,這樣就輕鬆實現了三種模式配置文件和平相處的局面。框架

1>.core-site.xml 配置文件dom

複製代碼
[root@cdh14 ~]$ more /tosp/opt/hadoop/etc/hadoop/core-site.xml 
<?xml version="1.0" encoding="UTF-8"?>
<configuration>
                <property>
                        <name>fs.defaultFS</name>
                        <value>hdfs://cdh14:9000</value>
                </property>
                <property>
                        <name>hadoop.tmp.dir</name>
                        <value>/tosp/opt/hadoop</value>
                </property>
</configuration>

<!--

core-site.xml配置文件的做用:
    用於定義系統級別的參數,如HDFS URL、Hadoop的臨時
目錄以及用於rack-aware集羣中的配置文件的配置等,此中的參
數定義會覆蓋core-default.xml文件中的默認配置。

fs.defaultFS 參數的做用:
        #聲明namenode的地址,至關於聲明hdfs文件系統。

hadoop.tmp.dir 參數的做用:
        #聲明hadoop工做目錄的地址。

-->
[root@cdh14 ~]$ 
複製代碼

2>.hdfs-site.xml 配置文件ssh

複製代碼
[root@cdh14 ~]$ more /tosp/opt/hadoop/etc/hadoop/hdfs-site.xml 
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>
        <property>
                <name>dfs.replication</name>
                <value>2</value>
        </property>
</configuration>

<!--
hdfs-site.xml 配置文件的做用:
        #HDFS的相關設定,如文件副本的個數、塊大小及是否使用強制權限
等,此中的參數定義會覆蓋hdfs-default.xml文件中的默認配置.

dfs.replication 參數的做用:
        #爲了數據可用性及冗餘的目的,HDFS會在多個節點上保存同一個數據
塊的多個副本,其默認爲3個。而只有一個節點的僞分佈式環境中其僅用
保存一個副本便可,這能夠經過dfs.replication屬性進行定義。它是一個
軟件級備份。

-->
[root@cdh14 ~]$ 
複製代碼

3>.mapred-site.xml 配置文件分佈式

複製代碼
[root@cdh14 ~]$ more /tosp/opt/hadoop/etc/hadoop/mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>
        <property>
                <name>mapreduce.framework.name</name>
                <value>yarn</value>
        </property>
</configuration>

<!--
mapred-site.xml 配置文件的做用:
        #HDFS的相關設定,如reduce任務的默認個數、任務所可以使用內存
的默認上下限等,此中的參數定義會覆蓋mapred-default.xml文件中的
默認配置.

mapreduce.framework.name 參數的做用:
        #指定MapReduce的計算框架,有三種可選,第一種:local(本地),第
二種是classic(hadoop一代執行框架),第三種是yarn(二代執行框架),我
們這裏配置用目前版本最新的計算框架yarn便可。

-->
[root@cdh14 ~]$ 
複製代碼

4>.yarn-site.xml配置文件ide

複製代碼
[root@cdh14 ~]$ more /tosp/opt/hadoop/etc/hadoop/yarn-site.xml 
<?xml version="1.0"?>
<configuration>
                <property>
                        <name>yarn.resourcemanager.hostname</name>
                        <value>cdh14</value>
                </property>
        <property>
                <name>yarn.nodemanager.aux-services</name>
                <value>mapreduce_shuffle</value>
        </property>
</configuration>

<!--

yarn-site.xml配置文件的做用:
        #主要用於配置調度器級別的參數.
yarn.resourcemanager.hostname 參數的做用:
        #指定資源管理器(resourcemanager)的主機名
yarn.nodemanager.aux-services 參數的做用:
        #指定nodemanager使用shuffle

-->
[root@cdh14 ~]$ 
複製代碼

 5>.slaves配置文件oop

[root@cdh14 ~]$ more /tosp/opt/hadoop/etc/hadoop/slaves 
#該配置文件的做用:是NameNode用與記錄須要鏈接哪些DataNode服務器節點,用與啓動或中止服務時發送遠程命令指令的目標主機。
cdh14
cdh12
cdh11
cdh10
cdh9
cdh8
cdh7 [root@cdh14 ~]$

 

三.在NameNode節點上配置免密碼登陸各DataNode節點spa

1>.在本地上生成公私祕鑰對(生成以前,把上次部署僞分佈式的祕鑰刪除掉)

複製代碼
[root@cdh14 ~]$ rm -rf ~/.ssh/*
[root@cdh14 ~]$ ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa
Generating public/private rsa key pair.
Your identification has been saved in /home/root/.ssh/id_rsa.
Your public key has been saved in /home/root/.ssh/id_rsa.pub.
The key fingerprint is:
a3:a4:ae:d8:f7:7f:a2:b6:d6:15:74:29:de:fb:14:08 root@cdh14
The key's randomart image is:
+--[ RSA 2048]----+
|             .   |
|          E o    |
|         o = .   |
|          o o .  |
|      . S  . . . |
|     o . .. . .  |
|    . .. .   o   |
| o .. o o .   .  |
|. oo.+++.o       |
+-----------------+
[root@cdh14 ~]$ 
複製代碼

2>.使用ssh-copy-id命令分配公鑰到DataNode服務器(172.20.20.228)

複製代碼
[root@cdh14 ~]$ ssh-copy-id root@cdh14
The authenticity of host 'cdh14 (172.16.30.101)' can't be established.
ECDSA key fingerprint is fa:25:bc:03:7e:99:eb:12:1e:bc:a8:c9:ce:39:ba:7b.
Are you sure you want to continue connecting (yes/no)? yes
/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed
/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed -- if you are prompted now it is to install the new keys
root@cdh14's password: 

Number of key(s) added: 1

Now try logging into the machine, with:   "ssh 'root@cdh14'"
and check to make sure that only the key(s) you wanted were added.

[root@cdh14 ~]$ ssh cdh14
Last login: Fri May 25 18:35:40 2018 from 172.16.30.1
[root@cdh14 ~]$ who
root pts/0        2018-05-25 18:35 (172.16.30.1)
root pts/1        2018-05-25 19:17 (cdh14)
[root@cdh14 ~]$ exit 
logout
Connection to cdh14 closed.
[root@cdh14 ~]$ who
root pts/0        2018-05-25 18:35 (172.16.30.1)
[root@cdh14 ~]$ 
複製代碼

3>.使用ssh-copy-id命令分配公鑰到DataNode服務器(172.20.20.226-220)

複製代碼
[root@cdh14 ~]$ ssh-copy-id root@chd12-cdh7
/usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed
/usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed -- if you are prompted now it is to install the new keys
root@s102's password: 

Number of key(s) added: 1

Now try logging into the machine, with:   "ssh 'root@s102'"
and check to make sure that only the key(s) you wanted were added.

[root@cdh14 ~]$ ssh s102
Last login: Fri May 25 18:35:42 2018 from 172.16.30.1
[root@s102 ~]$ who
root pts/0        2018-05-25 18:35 (172.16.30.1)
root pts/1        2018-05-25 19:19 (cdh14)
[root@s102 ~]$ exit 
logout
Connection to s102 closed.
[root@cdh14 ~]$ who
root pts/0        2018-05-25 18:35 (172.16.30.1)
[root@cdh14 ~]$ 
複製代碼

 

  注意:以上是普通使配置免密登陸,root用戶配置方法一致,最好也配置上root用戶的免密登陸,由於下文我會執行相應的shell腳本。

五.啓動服務並驗證是否成功

1>.格式化文件系統

root@cdh14 ~]$ hdfs namenode -format

2>.啓動hadoop

[root@cdh14 ~]$ start-all.sh

3>.用自定義腳本驗證NameNode和DataNode是否已經正常啓動

[root@cdh14 ~]$ jps
相關文章
相關標籤/搜索