ubuntu18.04.2 hadoop3.1.2+zookeeper3.5.5高可用完全分布式集群搭建


ubuntu18.04.2 hadoop3.1.2+zookeeper3.5.5高可用完全分布式集群搭建

集群規划:

hostname NameNode DataNode JournalNode ResourceManager Zookeeper
node01
node02
node03
node04
node05

准備工作:

首先克隆5台ubuntu虛擬機

1568421995422

vim /etc/netplan/01-network-manager-all.yaml修改網絡配置

我的5台網絡配置如下: (ps: 由於這次是家里台式, 所以網關和之前筆記本搭的那次不一樣)

# Let NetworkManager manage all devices on this system
# node01
network:
  version: 2
  renderer: NetworkManager
  ethernets:
    ens33:
        dhcp4: no
        dhcp6: no
        addresses: [192.168.180.130/24]
        gateway4: 192.168.180.2
        nameservers:
            addresses: [114.114.114.114, 8.8.8.8]
            
# Let NetworkManager manage all devices on this system
# node02
network:
  version: 2
  renderer: NetworkManager
  ethernets:
    ens33:
        dhcp4: no
        dhcp6: no
        addresses: [192.168.180.131/24]
        gateway4: 192.168.180.2
        nameservers:
            addresses: [114.114.114.114, 8.8.8.8]

# Let NetworkManager manage all devices on this system
# node03
network:
  version: 2
  renderer: NetworkManager
  ethernets:
    ens33:
        dhcp4: no
        dhcp6: no
        addresses: [192.168.180.132/24]
        gateway4: 192.168.180.2
        nameservers:
            addresses: [114.114.114.114, 8.8.8.8]
# Let NetworkManager manage all devices on this system
network:
  version: 2
  renderer: NetworkManager

  ethernets:
    ens33:
        dhcp4: no
        dhcp6: no
        addresses: [192.168.180.133/24]
        gateway4: 192.168.180.2
        nameservers:
            addresses: [114.114.114.114, 8.8.8.8]
# Let NetworkManager manage all devices on this system
network:
  version: 2
  renderer: NetworkManager

  ethernets:
    ens33:
        dhcp4: no
        dhcp6: no
        addresses: [192.168.180.134/24]
        gateway4: 192.168.180.2
        nameservers:
            addresses: [114.114.114.114, 8.8.8.8]

修改完畢后netplan apply 應用該網絡配置, ping一下百度有連接網絡配置就ok了。

修改主機名

vim /etc/hostname 將對應主機名分別修改為node01, niode02, node03, node04, node05

修改hosts文件

vim /etc/hosts 修改每台機器的hosts文件為如下:

127.0.0.1       localhost
192.168.180.130       node01
192.168.180.131       node02
192.168.180.132       node03
192.168.180.133       node04
192.168.180.134       node05

# The following lines are desirable for IPv6 capable hosts
::1     ip6-localhost ip6-loopback
fe00::0 ip6-localnet
ff00::0 ip6-mcastprefix
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters

JDK配置

如果沒有配置jdk請參考https://www.cnblogs.com/ronnieyuan/p/11461377.html

如果之前有下過別的版本的jdk請將jdk的tar包解壓到/usr/lib/jvm, 比如我的jvm目錄為如下:

drwxr-xr-x   4 root root 4096 9月  13 08:57 ./
drwxr-xr-x 133 root root 4096 9月  13 08:57 ../
lrwxrwxrwx   1 root root   25 4月   8  2018 default-java -> java-1.11.0-openjdk-amd64/
lrwxrwxrwx   1 root root   21 3月  27 04:57 java-1.11.0-openjdk-amd64 -> java-11-openjdk-amd64/
-rw-r--r--   1 root root 1994 3月  27 04:57 .java-1.11.0-openjdk-amd64.jinfo
drwxr-xr-x   9 root root 4096 4月  25 20:43 java-11-openjdk-amd64/
drwxr-xr-x   7 uucp  143 4096 12月 16  2018 jdk1.8/

然后再修改配置文件並選擇jdk(指令在參考中已寫)

jdk版本信息:

root@node01:~# java -version
java version "1.8.0_202"
Java(TM) SE Runtime Environment (build 1.8.0_202-b08)
Java HotSpot(TM) 64-Bit Server VM (build 25.202-b08, mixed mode)

免密登錄

在每台機器上執行ssh-keygen -t rsa -P "" (記得enter)

示例

root@node01:~# ssh-keygen -t rsa -P ""
Generating public/private rsa key pair.
Enter file in which to save the key (/root/.ssh/id_rsa): 
Created directory '/root/.ssh'.
Your identification has been saved in /root/.ssh/id_rsa.
Your public key has been saved in /root/.ssh/id_rsa.pub.
The key fingerprint is:
SHA256:el2s+e9UXxWjfGY1LS6RYD1CcHLmlXY+zJCopqRnuf0 root@node01
The key's randomart image is:
+---[RSA 2048]----+
|        ooB+.+ +o|
|         Bo.@ + *|
|         ..o % =.|
|      . o  .. X .|
|     o +S   o. .o|
|    . =. . +   .o|
|     o.o. +   . .|
|      ...  . .   |
|         .E .oo  |
+----[SHA256]-----+

在~目錄下vim .ssh/authorized_keys:

將5台虛擬機公鑰都存入該文件中, 每台的authorized_keys都一致

ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDBw4yPomSFt009LQ3gvxv9vnAF4tSXrJvVBMkpoi78mLMspgxYW6q3vLCWFEHT6HOLrLAQ/+UjclXjuVEEUGVOyn+dgvX7fK+XCOuTVdTyJZ3nIGbHUZ5zB+KHcJN3tiGjFQ3vGEuUeVkQ4jkN5RXI33nSx1eUM/sOuXtQ7DdhJjAuBko7RNw/jjTXW8znv8l8n5hb4fu4B+2CLkIkO+1+mTu8hljE2B+pu4o6cIiY/RTb0hNRLSs6w7K7BJFa+3ZkeMtiLtI8MUaIQzo4/nv4FKa8/GSvxLyyBZGoaunAYsUn7qmlNxNjEXY7wojAnVkRMiyCsEXQU6cEsR//Zocz root@node01

ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDg2AsPvf9TjjIVUlZutDqxFH579THtl6e7/SxYxHJ/def/T4dY5glzwW3AJ30Gcsw+k9E8PKiZIAiaQ7kU4/EmFK9LFhAuQx+glZS5GS88lXv7qSYOLmZtJPp0l4tgrIgk9u+PtZToCdlWpGLO2Xi3Dfggt//Lsl4Dqhl3dtrpZSjMGY7zkAd4fu696ri4rjv3kDciUdFNlKBFBkGA4RNFKylkPTlxLZfpqNU2pkZtBySHsGbEHMvnMQ1KOXRoW7pVvZ4QveR/eiQVXqq+v53oZ5KUmC5jpp6Abe3PVa7tG6s2ZOSP9ikOuFKrwXWArjp5H4oaYZIF/UenhhIdjxh3 root@node02

ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDwtqNeAwYLWqY5otArMcKd4iMBCpZ5cd+RyECunVnmeefuN1U53fR+h2UcS6/Jr84ZKlDmJ5+r9jgcBPIftbkGi9RE4aHEqo14sC3P4t6DODxNCI+enytx5/kw3gpKmxOdanrtojSWLdL+5v/h4qPt5e8AFfxqJ9HfZ5darXgRLWbkYcBADH51XvisY9Gf+DJKPjcD+3E8gMbHHdeYWt0crOkxbRVgnjmZVuWsYBRFH5x6ueR5SOHUC3WPzfeEdBvIeRddl4y1DvtvZZuVOxs1rQF59KdDSKSKt4s1lScZS1Kc57yXY2s+L6HrFqxfOO0u1pisfiDwDKvZDwKeMd3n root@node03

ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC+QfivNznStFt8xCZ1Qav6jKdErir0VbNRN0nqJaXUe+KL8YmYygofKEZRGQHCpYY2/rM7Cla6Pl9HLoatbvi89OYVy7V3hnu7SJwrqbkAGOqxCzW+OGdV9GRvhi3LTwJMAKxSrXB73tKK9ZqJd7WrP7o7ibyYMAbUiJTc0qa4gSXxXTunUuF2hOG7D88/93bxXXqSI9AydWrXBVxzmrP7CipXFOBqVC/mA/8SEdbVxSK0oGwa9KAAm690onoVevOVtTXWcvKSE/57WM94KJMbIKM/ypxKtUqKrgKuMfBsgs31Zu1j3SDkFC3Vm8uGj4yKnpxsaVJOwuMoRYiW90tT root@node04

ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDF/mqbRAwPusxpz5FA9FtIa97QJSuXjaRP+/37S7JvtCAh2FvgPBLIQeAdp7hvc/RFJ8WqDlQWj2UVpBsu2sn3Kg2VZ30qEghMLkMcCTtKknNX+U7SvBWCRoGojxl9lmi/Y1kkVNQUTRPQ8QeNGN2SvUi5A4Q+X1H6MEy16sLuamMlXqiIeqttY33odXj6oXI6OFqoE98FrNbTBrPwJFCk4Uhgnplbb0YE+4dbs9mVdR/iHpGm84WfvITe6Rn9Ry4K+Wo4C+Bms4dGfcO8eh8lrwSCff2IUIc877Zzc6ImYrdvZu7rvrCPyfNdoCJzA5wtExPoAfUbuN5T77ieLgWH root@node05

依次測試免密登錄是否成功

成功案例:

root@node01:~# ssh root@192.168.180.131
The authenticity of host '192.168.180.131 (192.168.180.131)' can't be established.
ECDSA key fingerprint is SHA256:++PMZ5boD2CgToi43EdaCSLtNGdVFt0xxCBoAIkggqk.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '192.168.180.131' (ECDSA) to the list of known hosts.
Welcome to Ubuntu 18.04.2 LTS (GNU/Linux 4.18.0-17-generic x86_64)

 * Documentation:  https://help.ubuntu.com
 * Management:     https://landscape.canonical.com
 * Support:        https://ubuntu.com/advantage


 * Canonical Livepatch is available for installation.
   - Reduce system reboots and improve kernel security. Activate at:
     https://ubuntu.com/livepatch

254 packages can be updated.
253 updates are security updates.

Your Hardware Enablement Stack (HWE) is supported until April 2023.
Last login: Sat Sep 14 08:44:54 2019 from 192.168.180.1
root@node02:~# 

Zookeeper3.5.5安裝

注:zookeeper只安裝在node03, node04 和node05上

上傳zookeeper3.5.5到/home/ronnie/soft目錄下

root@node03:/home/ronnie/soft# ll
total 524532
drwxr-xr-x  2 root   root        4096 9月  14 09:51 ./
drwxr-xr-x 32 ronnie ronnie      4096 9月  14 08:39 ../
-rw-r--r--  1 root   root    10622522 9月  13 11:35 apache-zookeeper-3.5.5-bin.tar.gz
-rw-r--r--  1 root   root   332433589 9月  13 09:18 hadoop-3.1.2.tar.gz
-rw-r--r--  1 root   root   194042837 1月  18  2019 jdk-8u202-linux-x64.tar.gz

tar -zxvf apache-zookeeper-3.5.5-bin.tar.gz -C /opt/ronnie/ 將其解壓至/opt/ronnie目錄

修改zookeeper目錄名:

cd /opt/ronnie
mv apache-zookeeper-3.5.5-bin/ zookeeper

創建並修改zookeeper配置文件

首先進入配置文件目錄:

root@node03:/opt/ronnie# cd zookeeper/conf/
root@node03:/opt/ronnie/zookeeper/conf# ll
total 20
drwxr-xr-x 2 2002 2002 4096 4月   2 21:05 ./
drwxr-xr-x 6 root root 4096 9月  14 09:54 ../
-rw-r--r-- 1 2002 2002  535 2月  15  2019 configuration.xsl
-rw-r--r-- 1 2002 2002 2712 4月   2 21:05 log4j.properties
-rw-r--r-- 1 2002 2002  922 2月  15  2019 zoo_sample.cfg

拷貝一份zoo_sample.cfg 為 zoo.cfg:

cp zoo_sample.cfg zoo.cfg
vim zoo.cfg
# The number of milliseconds of each tick
tickTime=2000
# The number of ticks that the initial 
# synchronization phase can take
initLimit=10
# The number of ticks that can pass between 
# sending a request and getting an acknowledgement
syncLimit=5
# the directory where the snapshot is stored.
# do not use /tmp for storage, /tmp here is just 
# example sakes.
dataDir=/var/ronnie/zookeeper
# the port at which the clients will connect
clientPort=2181
# the maximum number of client connections.
# increase this if you need to handle more clients
#maxClientCnxns=60
#
# Be sure to read the maintenance section of the 
# administrator guide before turning on autopurge.
#
# http://zookeeper.apache.org/doc/current/zookeeperAdmin.html#sc_maintenance
#
# The number of snapshots to retain in dataDir
#autopurge.snapRetainCount=3
# Purge task interval in hours
# Set to "0" to disable auto purge feature
#autopurge.purgeInterval=1
server.1=node03:2888:3888
server.2=node04:2888:3888
server.3=node05:2888:3888

將配置好的zookeeper目錄傳送給其他2台機子

scp -r /opt/ronnie/zookeeper/ root@192.168.180.133:/opt/ronnie/
scp -r /opt/ronnie/zookeeper/ root@192.168.180.134:/opt/ronnie/

在Datadir下創建myid, 並分別指定1, 2, 3

node03上的操作(其他同理), 若沒有該目錄需要先創建mkdir -p /var/ronnie/zookeeper/

cd /var/ronnie/zookeeper/
touch myid
echo 1 > myid

啟動Zookeeper

/opt/ronnie/zookeeper/bin/zkServer.sh start
# 若啟動成功
ZooKeeper JMX enabled by default
Using config: /opt/ronnie/zookeeper/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED

# 檢測狀態
/opt/ronnie/zookeeper/bin/zkServer.sh status

# 這是一個從節點
ZooKeeper JMX enabled by default
Using config: /opt/ronnie/zookeeper/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost.
Mode: follower

# 這是一個主節點
ZooKeeper JMX enabled by default
Using config: /opt/ronnie/zookeeper/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost.
Mode: leader


關閉zookeeper:

/opt/ronnie/zookeeper/bin/zkServer.sh stop

那么zookeeper到這里就安裝完畢了

Hadoop配置

vim ~/.bashrc 添加hadoop路徑

#HADOOP VARIABLES
export HADOOP_HOME=/opt/ronnie/hadoop-3.1.2
export PATH=$PATH:$HADOOP_HOME/bin
export PATH=$PATH:$HADOOP_HOME/sbin
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export YARN_HOME=$HADOOP_HOME
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib"

source ~/.bashrc 使新配置生效(記得每台都要改)

hadoop version查看版本, 顯示如下則hadoop路徑配置成功:

Hadoop 3.1.2
Source code repository https://github.com/apache/hadoop.git -r 1019dde65bcf12e05ef48ac71e84550d589e5d9a
Compiled by sunilg on 2019-01-29T01:39Z
Compiled with protoc 2.5.0
From source with checksum 64b8bdd4ca6e77cce75a93eb09ab2a9
This command was run using /opt/ronnie/hadoop-3.1.2/share/hadoop/common/hadoop-common-3.1.2.jar

修改hadoop-env.sh, mapred-env.sh, yarn-env.sh中的JAVA_HOME路徑, 沒有就在底下添加

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/hadoop-env.sh

 53 # variable is REQUIRED on ALL platforms except OS X!
 54 export JAVA_HOME=/usr/lib/jvm/jdk1.8

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/mapred-env.sh

 47 # JDK
 48 export JAVA_HOME=/usr/lib/jvm/jdk1.8

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/yarn-env.sh

171 # JDK
172 export JAVA_HOME=/usr/lib/jvm/jdk1.8

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/core-site.xml配置core-site.xml文件

  1 <?xml version="1.0" encoding="UTF-8"?>
  4   Licensed under the Apache License, Version 2.0 (the "License");
  7 
  8     http://www.apache.org/licenses/LICENSE-2.0
  9 
 12   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 13   See the License for the specific language governing permissions and
 14   limitations under the License. See accompanying LICENSE file.
 15 -->
 16 
 17 <!-- Put site-specific property overrides in this file. -->
 18 
 19 <configuration>
 20         <!-- 指定hdfs的nameservice名稱 -->
 21         <property>
 22                 <name>fs.defaultFS</name>
 23                 <value>hdfs://ns</value>
 24         </property>
 25         <!-- 指定臨時目錄 -->
 26         <property>
 27                 <name>hadoop.tmp.dir</name>
 28                 <value>/var/ronnie/hadoop/tmp</value>
 29         </property>
 30         <!-- 指定zookeeper  -->
 31         <property>
 32                 <name>ha.zookeeper.quorum</name>
 33                 <value>node03:2181,node04:2181,node05:2181</value>
 34         </property>
 35         <!-- Namenode向JournalNode發起的ipc連接請求重試最大次數  -->
 36         <property>
 37                 <name>ipc.client.connect.max.retries</name>
 38                 <value>100</value>
 39                 <description>Indicates the number of retries a client will make to establish a server c    onnection.
 40                 </description>
 41         </property>
 42         <!-- Namenode向JournalNode發起的ipc連接請求的重試間隔時間 -->
 43         <property>
 44                 <name>ipc.client.connect.retry.interval</name>
 45                 <value>10000</value>
 46                 <description>Indicates the number of milliseconds a client will wait for before retryin    g to establish.
 47                 </description>
 48         </property>
 49         <!-- 開啟回收功能, 並設置垃圾刪除間隔(min)  -->
 50         <property>
 51                 <name>fs.trash.interval</name>
 52                 <value>360</value>
 53                 <description>
 54                         Trash deletion interval in minutes. If zero, the trash feature is diabled.
 55                 </description>
 56         </property>
 57         <!-- 設置垃圾檢查點介個(min), 不設置的話默認和fs.trash.interval一樣  -->
 58         <property>
 59                 <name>fs.trash.checkpoint.interval</name>
 60                 <value>60</value>
 61                 <description>
 62                         Trash checkpoint interval in minutes. If zero, the deletion interval is used.
 63                 </description>
 64         </property>
 65         <!-- 配置oozie時使用以下參數 -->
 66         <property>
 67                 <name>hadoop.proxyuser.deplab.groups</name>
 68                 <value>*</value>
 69         </property>
 70         <property>
 71                 <name>hadoop.proxyuser.deplab.hosts</name>
 72                 <value>*</value>
 73         </property>
 74 </configuration>

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/hdfs-site.xml修改hdfs-site.xml

  1 <?xml version="1.0" encoding="UTF-8"?>
  2 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3 <!--
  4   Licensed under the Apache License, Version 2.0 (the "License");
  5   you may not use this file except in compliance with the License.
  6   You may obtain a copy of the License at
  7 
  8     http://www.apache.org/licenses/LICENSE-2.0
  9 
 10   Unless required by applicable law or agreed to in writing, software
 11   distributed under the License is distributed on an "AS IS" BASIS,
 12   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 13   See the License for the specific language governing permissions and
 14   limitations under the License. See accompanying LICENSE file.
 15 -->
 16 
 17 <!-- Put site-specific property overrides in this file. -->
 18 
 19 <configuration>
 20         <!-- 指定hdfs的nameservice為ns, 需要和core-site.xml中的保持一致  -->
 21         <property>
 22                 <name>dfs.nameservices</name>
 23                 <value>ns</value>
 24         </property>
 25         <!-- ns下有nn1, nn2 兩個NameNode  -->
 28                 <value>nn1,nn2</value>
 29         </property>
 30         <!-- nn1的RPC通信地址 -->
 33                 <value>node01:9000</value>
 34         </property>
 35         <!-- nn1 的http通信地址 -->
 36         <property>
 37                 <name>dfs.namenode.http-address.ns.nn1</name>
 38                 <value>node01:50070</value>
 39         </property>
 40         <!-- nn2的RPC通信地址 -->
 41         <property>
 42                 <name>dfs.namenode.rpc-address.ns.nn2</name>
 43                 <value>node02:9000</value>
 44         </property>
 45         <!-- nn2 的http通信地址 -->
 46         <property>
 47                 <name>dfs.namenode.http-address.ns.nn2</name>
 48                 <value>node02:50070</value>
 49         </property>
 50         <!-- 指定NameNode的edits元數據在JournalNode上的存放位置 -->
 51         <property>
 52                 <name>dfs.namenode.shared.edits.dir</name>
 53                 <value>qjournal://node03:8485;node04:8485;node05:8485/ns</value>
 54         </property>
 55         <!-- 指定JournalNode在本地磁盤存放數據的尾椎  -->
 56         <property>
 57                 <name>dfs.journalnode.edits.dir</name>
 58                 <value>/var/ronnie/hadoop/jdata</value>
 59         </property>
 60         <!-- 開啟NameNode失敗自動切換 -->
 61         <property>
 62                 <name>dfs.ha.automatic-failover.enabled</name>
 63                 <value>true</value>
 64         </property>
 65         <!-- 配置失敗自動切換實現方式  -->
 66         <property>
 67                 <name>dfs.client.failover.proxy.provider.ns</name>
 68                 <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value    >
 69         </property>
 70         <!-- 配置隔離機制方法, 多個機制用換行分割, 即每個機制暫用一行  -->
 71         <property>
 72                 <name>dfs.ha.fencing.methods</name>
 73                 <value>
 74                         sshfence
 75                         shell(/bin/true)
 76                 </value>
 77         </property>
 78         <!-- 使用sshfence隔離機制時需要ssh免密登錄 -->
 79         <property>
 80                 <name>dfs.ha.fencing.ssh.private-key-files</name>
 81                 <value>/root/.ssh/id_rsa</value>
 82         </property>
 83         <!-- 配置sshfence隔離機制超時時間  -->
 84         <property>
 85                 <name>dfs.ha.fencing.ssh.connect-timeout</name>
 86                 <value>30000</value>
 87         </property>
 88 </configuration>

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/mapred-site.xml修改mapred-site.xml

  1 <?xml version="1.0"?>
  2 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3 <!--
  4   Licensed under the Apache License, Version 2.0 (the "License");
  5   you may not use this file except in compliance with the License.
  6   You may obtain a copy of the License at
  7 
  8     http://www.apache.org/licenses/LICENSE-2.0
  9 
 10   Unless required by applicable law or agreed to in writing, software
 11   distributed under the License is distributed on an "AS IS" BASIS,
 12   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 13   See the License for the specific language governing permissions and
 14   limitations under the License. See accompanying LICENSE file.
 15 -->
 16 
 17 <!-- Put site-specific property overrides in this file. -->
 18 
 19 <configuration>
 20         <!-- 指定mr框架為yarn方式  -->
 21         <property>
 22                 <name>mapreduce.framework.name</name>
 23                 <value>yarn</value>
 24         </property>
 25 </configuration>

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/yarn-site.xml修改yarn-site.xml

  1 <?xml version="1.0"?>
  2 <!--
  3   Licensed under the Apache License, Version 2.0 (the "License");
  4   you may not use this file except in compliance with the License.
  5   You may obtain a copy of the License at
  6 
  7     http://www.apache.org/licenses/LICENSE-2.0
  8 
  9   Unless required by applicable law or agreed to in writing, software
 10   distributed under the License is distributed on an "AS IS" BASIS,
 11   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 12   See the License for the specific language governing permissions and
 13   limitations under the License. See accompanying LICENSE file.
 14 -->
 15 <configuration>
 16         <!-- 開啟RM高可用  -->
 17         <property>
 18                 <name>yarn.resourcemanager.ha.enabled</name>
 19                 <value>true</value>
 20         </property>
 21         <!-- 指定RM的集群id  -->
 22         <property>
 23                 <name>yarn.resourcemanager.cluster-id</name>
 24                 <value>yrc</value>
 25         </property>
 26         <!-- 指定RM的名稱  -->
 27         <property>
 28                 <name>yarn.resourcemanager.ha.rm-ids</name>
 29                 <value>rm1,rm2</value>
 30         </property>
 31         <!-- 指定rm1, rm2的地址  -->
 32         <property>
 33                 <name>yarn.resourcemanager.hostname.rm1</name>
 34                 <value>node01</value>
 35         </property>
 36         <property>
 37                 <name>yarn.resourcemanager.hostname.rm2</name>
 38                 <value>node02</value>
 39         </property>
 40         <!-- 指定zookeeper集群地址  -->
 41         <property>
 42                 <name>yarn.resourcemanager.zk-address</name>
 43                 <value>node03:2181,node04:2181,node05:2181</value>
 44         </property>
 45         <!-- 設定洗牌  -->
 46         <property>
 47                 <name>yarn.nodemanager.aux-services</name>
 48                 <value>mapreduce_shuffle</value>
 49         </property>
 50 </configuration>

vim /opt/ronnie/hadoop-3.1.2/etc/hadoop/workers修改工作組

node01
node02
node03
node04
node05

vim /opt/ronnie/hadoop-3.1.2/sbin/start-dfs.sh

vim /opt/ronnie/hadoop-3.1.2/sbin/stop-dfs.sh

在文件頂部頭文件之后添加:

#!/usr/bin/env bash
  
HDFS_DATANODE_USER=root

HADOOP_SECURE_DN_USER=hdfs

HDFS_NAMENODE_USER=root

HDFS_SECONDARYNAMENODE_USER=root

HDFS_JOURNALNODE_USER=root

HDFS_ZKFC_USER=root

vim /opt/ronnie/hadoop-3.1.2/sbin/start-yarn.sh

vim /opt/ronnie/hadoop-3.1.2/sbin/stop-yarn.sh

在文件頂部頭文件之后添加:

#!/usr/bin/env bash
  
YARN_RESOURCEMANAGER_USER=root

HADOOP_SECURE_DN_USER=yarn

YARN_NODEMANAGER_USER=root

將配置好的hadoop目錄傳送給其他節點

scp -r /opt/ronnie/hadoop-3.1.2/ root@192.168.180.131:/opt/ronnie/
scp -r /opt/ronnie/hadoop-3.1.2/ root@192.168.180.132:/opt/ronnie/
scp -r /opt/ronnie/hadoop-3.1.2/ root@192.168.180.133:/opt/ronnie/
scp -r /opt/ronnie/hadoop-3.1.2/ root@192.168.180.134:/opt/ronnie/

啟動集群

先啟動zookeeper集群:

root@node03:/var/ronnie/zookeeper# /opt/ronnie/zookeeper/bin/zkServer.sh start
root@node04:/var/ronnie/zookeeper# /opt/ronnie/zookeeper/bin/zkServer.sh start
root@node05:/var/ronnie/zookeeper# /opt/ronnie/zookeeper/bin/zkServer.sh start

分別在node03, node04, node05上啟動journalnode

root@node03:~# /opt/ronnie/hadoop-3.1.2/sbin/hadoop-daemon.sh start journalnode
root@node04:~# /opt/ronnie/hadoop-3.1.2/sbin/hadoop-daemon.sh start journalnode
root@node05:~# /opt/ronnie/hadoop-3.1.2/sbin/hadoop-daemon.sh start journalnode

jps查看進程

root@node03:~# jps
6770 Jps
6724 JournalNode
6616 QuorumPeerMain

在兩台NameNode中選一台進行格式化(這里選node01)

root@node01:~# hdfs namenode -format

如有報錯會寫明哪里配置文件寫錯錯了, 再回去改。

開啟NameNode:

hdfs --daemon start namenode

jps查看是否開啟

root@node01:~# jps
5622 Jps
5549 NameNode

在另一台NameNode上同步格式化后的相關信息

root@node02:~# hdfs namenode  -bootstrapStandby

在node01上格式化ZKFC(只需要執行一次)

root@node01:~# hdfs zkfc -formatZK

啟動hdfs集群

root@node01:~# start-dfs.sh

打開node01的50070端口

1568437495202

1568437470742

啟動yarn集群

root@node01:~# start-yarn.sh

打開node01的8088端口

1568437638912

到此配置成功
如果上傳文件遇到permission denied,則在hdfs-site.xml文件末尾添加:

        <property>
                <name>dfs.permissions</name>
                <value>false</value>
        </property>


免責聲明!

本站轉載的文章為個人學習借鑒使用,本站對版權不負任何法律責任。如果侵犯了您的隱私權益,請聯系本站郵箱yoyou2525@163.com刪除。



 
粵ICP備18138465號   © 2018-2025 CODEPRJ.COM