一、JDK安裝
JDK官網下載:https://www.oracle.com
1.創建軟件存放目錄
[root@hadoop100 ~]# mkdir /data/software/
[root@hadoop100 ~]# cd /data/software/
2.上傳JDK安裝包
[root@hadoop100 /data/software]# rz
[root@hadoop100 /data/software]# ll
total 206884
-rw-r--r--. 1 root root 185540433 2021-07-12 09:42 jdk-8u131-linux-x64.tar.gz
3.解壓安裝包
[root@hadoop100 /data/software]# tar xf jdk-8u131-linux-x64.tar.gz -C /opt
[root@hadoop100 /data/software]# cd /opt/
[root@hadooop100 ~]# ll
total 0
drwxr-xr-x. 8 10 143 255 2021-03-15 16:35 jdk1.8.0_131
4.做軟連接
[root@hadooop100 ~]# ln -s jdk1.8.0_131 jdk
[root@hadooop100 ~]# ll
total 0
lrwxrwxrwx. 1 root root 12 2021-07-12 09:43 jdk -> jdk1.8.0_131
drwxr-xr-x. 8 10 143 255 2021-03-15 16:35 jdk1.8.0_131
5.配置環境變量
[root@hadooop100 ~]# vim /etc/profile.d/java.sh
export JAVA_HOME=/opt/jdk
export PATH=$PATH:$JAVA_HOME/bin
6.刷新環境變量
[root@hadooop100 ~]# source /etc/profile.d/java.sh
7.驗證JDK
[root@hadoop100 /opt]# java -version
java version "1.8.0_131"
Java(TM) SE Runtime Environment (build 1.8.0_131-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.131-b11, mixed mode)
二、SSH配置免密登錄
#1.若沒有該目錄,可以創建
[root@hadoop100 ~/.ssh]# vim /etc/hosts
127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
68.79.16.69 hadoop100
[root@hadoop100 ~]# cd ~/.ssh/
#2.會有提示,一路按回車就可以
[root@hadoop100 ~/.ssh]# ssh-keygen -t rsa
Generating public/private rsa key pair.
Enter file in which to save the key (/root/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /root/.ssh/id_rsa.
Your public key has been saved in /root/.ssh/id_rsa.pub.
The key fingerprint is:
SHA256:ka+QA6wxsf+MdkstlnbIT2sVwqGV7q1lU9RNy2KKjDY root@hadoop100
The key's randomart image is:
+---[RSA 2048]----+
| . . . o.|
| + +. . o o|
| + o =o. . o o |
| = ...*oo + . |
| . . +ES=.+ |
| =.*o.* |
| o % == . |
| . = *o. |
| ..o |
+----[SHA256]-----+
#3.添加免密
[root@hadoop100 ~/.ssh]# cat ./id_rsa.pub >> ./authorized_keys
#4.測試免密登錄
[root@hadoop100 ~/.ssh]# ssh root@hadoop100
The authenticity of host 'hadoop100 (fe80::20c:29ff:fef5:98f9%eth0)' can't be established.
ECDSA key fingerprint is SHA256:g6buQ4QMSFl+5MMAh8dTCmLtkIfdT8sgRFYc6uCzV3c.
ECDSA key fingerprint is MD5:5f:d7:ad:07:e8:fe:d2:49:ec:79:2f:d4:91:59:c5:03.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added 'hadoop100,fe80::20c:29ff:fef5:98f9%eth0' (ECDSA) to the list of known hosts.
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
Last login: Mon Aug 30 11:24:23 2021 from ::1
[root@hadoop100 ~]#
三、Hadoop安裝
Hadoop官網下載地址:https://hadoop.apache.org/releases.html
1.下載安裝包
[root@hadooop100 ~]# cd /data/software/
[root@hadooop100 /data/software]# wget https://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/hadoop-3.3.1/hadoop-3.3.1.tar.gz
[root@hadooop100 /data/software]# ll
總用量 1398724
-rw-r--r-- 1 root root 605187279 6月 15 17:55 hadoop-3.3.1.tar.gz
2.解壓安裝包
[root@hadooop100 /data/software]# tar xf hadoop-3.3.1.tar.gz -C /opt/
[root@hadooop100 /data/software]# cd /opt/
[root@hadooop100 /opt]# ll
總用量 0
drwxr-xr-x 10 centos centos 215 6月 15 13:52 hadoop-3.3.1
3.做軟連接
[root@hadooop100 /opt]# ln -s hadoop-3.3.1 hadoop
[root@hadooop100 /opt]# ll
總用量 0
lrwxrwxrwx 1 root root 12 8月 13 15:01 hadoop -> hadoop-3.3.1
drwxr-xr-x 10 centos centos 215 6月 15 13:52 hadoop-3.3.1
4.配置環境變量
[root@hadooop100 /opt]# vim /etc/profile.d/hadoop.sh
export HADOOP_HOME=/opt/hadoop
export PATH=$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib:$HADOOP_HOME/lib/native"
5.刷新環境變量
[root@hadoop100 /opt]# source /etc/profile.d/hadoop.sh
6.驗證hadoop
[root@hadoop100 /opt]# hadoop version
Hadoop 3.3.1
Source code repository https://github.com/apache/hadoop.git -r a3b9c37a397ad4188041dd80621bdeefc46885f2
Compiled by ubuntu on 2021-06-15T05:13Z
Compiled with protoc 3.7.1
From source with checksum 88a4ddb2299aca054416d6b7f81ca55
This command was run using /opt/hadoop-3.3.1/share/hadoop/common/hadoop-common-3.3.1.jar
四、Hadoop目錄結構
1.查看Hadoop目錄結構
[root@hadoop100 /opt]# cd /opt/hadoop
[root@hadoop100 /opt/hadoop]# ll
總用量 112
drwxr-xr-x. 2 nexus nexus 4096 2021-06-15 13:52 bin
drwxr-xr-x. 3 nexus nexus 4096 2021-06-15 13:15 etc
drwxr-xr-x. 2 nexus nexus 4096 2021-06-15 13:52 include
drwxr-xr-x. 3 nexus nexus 4096 2021-06-15 13:52 lib
drwxr-xr-x. 4 nexus nexus 4096 2021-06-15 13:52 libexec
-rw-rw-r--. 1 nexus nexus 23450 2021-06-15 13:02 LICENSE-binary
drwxr-xr-x. 2 nexus nexus 4096 2021-06-15 13:52 licenses-binary
-rw-rw-r--. 1 nexus nexus 15217 2021-06-15 13:02 LICENSE.txt
-rw-rw-r--. 1 nexus nexus 29473 2021-06-15 13:02 NOTICE-binary
-rw-rw-r--. 1 nexus nexus 1541 2021-05-22 00:11 NOTICE.txt
-rw-rw-r--. 1 nexus nexus 175 2021-05-22 00:11 README.txt
drwxr-xr-x. 3 nexus nexus 4096 2021-06-15 13:15 sbin
drwxr-xr-x. 4 nexus nexus 4096 2021-06-15 14:18 share
2.重要目錄含義
1)bin目錄:存放對Hadoop相關服務(hdfs,yarn,mapred)進行操作的腳本
2)etc目錄:Hadoop的配置文件目錄,存放Hadoop的配置文件
3)lib目錄:存放Hadoop的本地庫(對數據進行壓縮解壓縮功能)
4)sbin目錄:存放啟動或停止Hadoop相關服務的腳本
5)share目錄:存放Hadoop的依賴jar包、文檔、和官方案例
五、Hadoop配置
1.備份Hadoop配置
配置Hadoop,共需要配置5個文件(另外可選地配置workers文件),均位於Hadoop安裝目錄下的"etc/hadoop/"子目錄下。首先進入到該目錄下:
[root@hadooop100 /opt]# cd /opt/hadoop/etc/hadoop/
[root@hadooop100 /opt/hadoop/etc/hadoop]# cp hadoop-env.sh hadoop-env.sh.bak
[root@hadooop100 /opt/hadoop/etc/hadoop]# cp core-site.xml core-site.xml.bak
[root@hadooop100 /opt/hadoop/etc/hadoop]# cp hdfs-site.xml hdfs-site.xml.bak
[root@hadooop100 /opt/hadoop/etc/hadoop]# cp mapred-site.xml mapred-site.xml.bak
[root@hadooop100 /opt/hadoop/etc/hadoop]# cp yarn-site.xml yarn-site.xml.bak
[root@hadoop100 /opt/hadoop/etc/hadoop]# cp workers workers.bak
2.修改Hadoop配置
[root@hadooop100 /opt/hadoop/etc/hadoop]# vim hadoop-env.sh
export JAVA_HOME=/opt/jdk
export HDFS_DATANODE_USER=root
[root@hadooop100 /opt/hadoop/etc/hadoop]# vim core-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://hadoop100:8020/</value>
<description>HDFS的URI,文件系統://namenode標識:端口號--></description>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/data/hadoop/tmp</value>
<description>namenode上本地的hadoop臨時文件夾</description>
</property>
<property>
<name>hadoop.http.staticuser.user</name>
<value>root</value>
<description>配置HDFS網頁登錄使用的靜態用戶為root</description>
</property>
<property>
<name>hadoop.proxyuser.root.hosts</name>
<value>*</value>
<description>配置root用戶允許通過代理訪問的主機節點</description>
</property>
<property>
<name>hadoop.proxyuser.root.groups</name>
<value>*</value>
<description>配置root用戶允許代理的用戶所屬組</description>
</property>
</configuration>
[root@hadooop100 /opt/hadoop/etc/hadoop]# vim hdfs-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>dfs.namenode.http-address</name>
<value>hadoop100:9870</value>
<description>nn web端訪問地址</description>
</property>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>hadoop100:9868</value>
<description> 2nn web端訪問地址</description>
</property>
<property>
<name>dfs.replication</name>
<value>1</value>
<description>副本個數,配置默認是3,應小於datanode機器數量</description>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/data/hadoop/dfs/name</value>
<description>存放hadoop的名稱節點namenode里的metadata</description>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/data/hadoop/dfs/data</value>
<description>存放HDFS文件系統數據文件的目錄</description>
</property>
<property>
<name>dfs.datanode.max.transfer.threads</name>
<value>4096</value>
<description>HDFS數據節點文件數量上限</description>
</property>
</configuration>
[root@hadooop100 /opt/hadoop/etc/hadoop]# vim mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
<description>指定MapReduce程序運行在Yarn上</description>
</property>
</configuration>
[root@hadooop100 /opt/hadoop/etc/hadoop]# vim yarn-site.xml
<?xml version="1.0"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
<description>指定MR走shuffle</description>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop100</value>
<description>指定ResourceManager的地址</description>
</property>
<property>
<name>yarn.nodemanager.env-whitelist</name>
<value>JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME,HADOOP_MAPRED_HOME</value>
<description>環境變量的繼承</description>
</property>
</configuration>
[root@hadoop100 /opt/hadoop/etc/hadoop]# vim workers
hadoop100
[root@hadooop100 ~]# cd /opt/hadoop/sbin/
# 1.對於start-dfs.sh和stop-dfs.sh文件,添加下列參數:
#!/usr/bin/env bash
HDFS_DATANODE_USER=root
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root
HDFS_SECONDARYNAMENODE_USER=root
# 1.對於start-yarn.sh和stop-yarn.sh文件,添加下列參數:
#!/usr/bin/env bash
YARN_RESOURCEMANAGER_USER=root
HADOOP_SECURE_DN_USER=yarn
YARN_NODEMANAGER_USER=root
六、創建配置文件目錄
[root@hadooop100 ~]# mkdir -p /data/hadoop/tmp
[root@hadooop100 ~]# mkdir -p /data/hadoop/hdfs
[root@hadooop100 ~]# mkdir -p /data/hadoop/dfs/date
[root@hadooop100 ~]# mkdir -p /data/hadoop/dfs/name
七、啟動Hadoop
1.格式化HDFS文件系統
格式化是對 HDFS這個分布式文件系統中的 DataNode 進行分塊,統計所有分塊后的初始元數據的存儲在namenode中。(如果服務器再次啟動,也需要進行這步,否則啟動可能會失敗)
[root@hadoop100 ~]# hdfs namenode -format
2.Hadoop上執行MR程序
#1.啟動HDFS
[root@hadoop100 ~]# start-dfs.sh
Starting namenodes on [hadoop100]
Last login: Mon Aug 30 11:24:37 CST 2021 from fe80::20c:29ff:fef5:98f9%eth0 on pts/1
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
hadoop100: Warning: Permanently added the ECDSA host key for IP address '10.0.0.100' to the list of known hosts.
Starting datanodes
Last login: Mon Aug 30 11:53:39 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
Starting secondary namenodes [hadoop100]
Last login: Mon Aug 30 11:53:42 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
#2.使用jps查看是否啟動成功
[root@hadoop100 ~]# jps
26594 Jps
26115 NameNode
26243 DataNode
26445 SecondaryNameNode
3.啟動yarn
#1.啟動yarn
[root@hadoop100 ~]# start-yarn.sh
Starting resourcemanager
Last login: Mon Aug 30 11:53:57 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
Starting nodemanagers
Last login: Mon Aug 30 11:54:47 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
#2.使用jps查看是否啟動成功
[root@hadoop100 ~]# jps
26849 NodeManager
26115 NameNode
26243 DataNode
26932 Jps
26728 ResourceManager
26445 SecondaryNameNode
八、測試Web登錄
#1.在本地windows上的C:\Windows\System32\drivers\etc下有個hosts文件,修改hosts文件,增加以下內容:
68.79.16.69 hadoop100
#2.查看HDFS和YARN的web界面
http://主機IP:9870 查看hdfs是否OK
http://主機IP:8088 查看yarn是否OK
本機URL:
http://68.79.16.69:9870/dfshealth.html#tab-overview
http://68.79.16.69:8088/cluster


九、Hadoop基本測試
1.新建小文件
[root@hadoop100 ~]# vim /data/software/1.txt
c 是世界上最好的語言!
java 是世界上最好的語言!
python 是世界上最好的語言!
go 是世界上最好的語言!
2.上傳文件到Hadoop
#1.新建上傳目錄
[root@hadoop100 ~]# hadoop fs -mkdir /input
#2.上傳小文件
[root@hadoop100 ~]# hadoop fs -put /data/software/1.txt /input
#3.上傳大文件
[root@hadoop100 ~]# hadoop fs -put /data/software/jdk-8u131-linux-x64.tar.gz /input
3.校驗文件一致性
#1.登錄HDFS的Web界面,
http://68.79.16.69:9870/dfshealth.html#tab-overview
點擊Utilities --> Browse the file system --> 選擇 Browse Directory為/ 這里可以看到我們之前新建的上傳目錄

#2.點擊input,可以看到我們上傳的兩個文件1.txt和jdk的安裝包

#3.點擊上傳的文件,查看文件信息,點擊Download可以正常下載,點擊Tail the file可以看到文件內容

十、配置歷史服務器
1.配置歷史服務器
為了查看程序的歷史運行情況,需要配置一下歷史服務器。完整步驟如下:
[root@hadoop100 /opt/hadoop/etc/hadoop]# vim mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
<description>指定MapReduce程序運行在Yarn上</description>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>hadoop100:10020</value>
<description>歷史服務器端地址</description>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>hadoop100:19888</value>
<description>歷史服務器web端地址</description>
</property>
</configuration>
2.啟動歷史服務器
#1.在hadoop100啟動歷史服務器
[root@hadoop100 /opt/hadoop/etc/hadoop]# mapred --daemon start historyserver
#2.查看歷史服務器是否啟動
[root@hadoop100 /opt/hadoop/etc/hadoop]# jps
26849 NodeManager
26115 NameNode
26243 DataNode
26728 ResourceManager
29017 JobHistoryServer
26445 SecondaryNameNode
29039 Jps
#3.查看JobHistory
http://hadoop100:19888/jobhistory

3. 配置日志的聚集
日志聚集概念:應用運行完成以后,將程序運行日志信息上傳到HDFS系統上。
日志聚集功能好處:可以方便的查看到程序運行詳情,方便開發調試。
注意:開啟日志聚集功能,需要重新啟動NodeManager 、ResourceManager和HistoryServer,完整配置如下:
[root@hadoop100 /opt/hadoop/etc/hadoop]# vim yarn-site.xml
<?xml version="1.0"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
<description>指定MR走shuffle</description>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop100</value>
<description>指定ResourceManager的地址</description>
</property>
<property>
<name>yarn.nodemanager.env-whitelist</name>
<value>JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME
,HADOOP_MAPRED_HOME</value>
<description>環境變量的繼承</description>
</property>
<property>
<name>yarn.log-aggregation-enable</name>
<value>true</value>
<description>開啟日志聚集功能</description>
</property>
<property>
<name>yarn.log.server.url</name>
<value>http://hadoop100:19888/jobhistory/logs</value>
<description>設置日志聚集服務器地址</description>
</property>
<property>
<name>yarn.log-aggregation.retain-seconds</name>
<value>604800</value>
<description>設置日志保留時間為7天</description>
</property>
</configuration>
4.重啟NodeManager 、ResourceManager和HistoryServer
#1.關閉NodeManager 、ResourceManager和HistoryServer
[root@hadoop100 /opt/hadoop/etc/hadoop]# stop-yarn.sh
Stopping nodemanagers
Last login: Mon Aug 30 11:54:49 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
Stopping resourcemanager
Last login: Mon Aug 30 15:09:21 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
[root@hadoop100 /opt/hadoop/etc/hadoop]# mapred --daemon stop historyserver
#2.啟動NodeManager 、ResourceManager和HistoryServer
[root@hadoop100 /opt/hadoop/etc/hadoop]# start-yarn.sh
Starting resourcemanager
Last login: Mon Aug 30 15:09:25 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
Starting nodemanagers
Last login: Mon Aug 30 15:09:59 CST 2021 on pts/0
Last failed login: Mon Aug 30 17:24:51 CST 2021 on tty1
There was 1 failed login attempt since the last successful login.
[root@hadoop100 /opt/hadoop/etc/hadoop]# mapred --daemon start historyserver
5.執行WordCount程序
#1.執行WordCount程序
[root@hadoop100 /opt/hadoop/etc/hadoop]# cd /opt/hadoop
[root@hadoop100 /opt/hadoop]# hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-3.3.1.jar wordcount /input /output
2021-08-30 15:22:57,111 INFO client.DefaultNoHARMFailoverProxyProvider: Connecting to ResourceManager at hadoop100/10.0.0.100:8032
2021-08-30 15:22:58,773 INFO mapreduce.JobResourceUploader: Disabling Erasure Coding for path: /tmp/hadoop-yarn/staging/root/.staging/job_1630308143240_0001
2021-08-30 15:22:59,587 INFO input.FileInputFormat: Total input files to process : 2
2021-08-30 15:23:00,569 INFO mapreduce.JobSubmitter: number of splits:2
2021-08-30 15:23:01,548 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1630308143240_0001
2021-08-30 15:23:01,548 INFO mapreduce.JobSubmitter: Executing with tokens: []
2021-08-30 15:23:02,222 INFO conf.Configuration: resource-types.xml not found
2021-08-30 15:23:02,223 INFO resource.ResourceUtils: Unable to find 'resource-types.xml'.
2021-08-30 15:23:03,506 INFO impl.YarnClientImpl: Submitted application application_1630308143240_0001
2021-08-30 15:23:03,743 INFO mapreduce.Job: The url to track the job: http://hadoop100:8088/proxy/application_1630308143240_0001/
2021-08-30 15:23:03,743 INFO mapreduce.Job: Running job: job_1630308143240_0001
2021-08-30 15:23:29,788 INFO mapreduce.Job: Job job_1630308143240_0001 running in uber mode : false
2021-08-30 15:23:29,790 INFO mapreduce.Job: map 0% reduce 0%
2021-08-30 15:23:55,714 INFO mapreduce.Job: map 50% reduce 0%
2021-08-30 15:24:04,985 INFO mapreduce.Job: map 52% reduce 0%
2021-08-30 15:24:11,140 INFO mapreduce.Job: map 54% reduce 0%
2021-08-30 15:24:17,249 INFO mapreduce.Job: map 56% reduce 0%
2021-08-30 15:24:28,507 INFO mapreduce.Job: map 56% reduce 17%
2021-08-30 15:24:35,597 INFO mapreduce.Job: map 57% reduce 17%
2021-08-30 15:24:41,779 INFO mapreduce.Job: map 59% reduce 17%
2021-08-30 15:24:53,962 INFO mapreduce.Job: map 60% reduce 17%
2021-08-30 15:25:00,080 INFO mapreduce.Job: map 63% reduce 17%
2021-08-30 15:25:06,230 INFO mapreduce.Job: map 66% reduce 17%
2021-08-30 15:25:12,421 INFO mapreduce.Job: map 67% reduce 17%
2021-08-30 15:25:18,564 INFO mapreduce.Job: map 70% reduce 17%
2021-08-30 15:25:24,668 INFO mapreduce.Job: map 71% reduce 17%
2021-08-30 15:25:30,757 INFO mapreduce.Job: map 72% reduce 17%
2021-08-30 15:25:36,898 INFO mapreduce.Job: map 74% reduce 17%
2021-08-30 15:25:49,063 INFO mapreduce.Job: map 75% reduce 17%
2021-08-30 15:25:55,215 INFO mapreduce.Job: map 77% reduce 17%
2021-08-30 15:26:06,378 INFO mapreduce.Job: map 79% reduce 17%
2021-08-30 15:26:13,472 INFO mapreduce.Job: map 80% reduce 17%
2021-08-30 15:26:19,560 INFO mapreduce.Job: map 81% reduce 17%
2021-08-30 15:26:25,686 INFO mapreduce.Job: map 83% reduce 17%
2021-08-30 15:26:37,848 INFO mapreduce.Job: map 85% reduce 17%
2021-08-30 15:26:43,944 INFO mapreduce.Job: map 87% reduce 17%
2021-08-30 15:26:50,050 INFO mapreduce.Job: map 91% reduce 17%
2021-08-30 15:26:56,122 INFO mapreduce.Job: map 95% reduce 17%
2021-08-30 15:27:02,390 INFO mapreduce.Job: map 99% reduce 17%
2021-08-30 15:27:04,436 INFO mapreduce.Job: map 100% reduce 17%
2021-08-30 15:27:13,258 INFO mapreduce.Job: map 100% reduce 73%
2021-08-30 15:27:19,624 INFO mapreduce.Job: map 100% reduce 87%
2021-08-30 15:27:24,733 INFO mapreduce.Job: map 100% reduce 100%
2021-08-30 15:27:27,763 INFO mapreduce.Job: Job job_1630308143240_0001 completed successfully
2021-08-30 15:27:28,727 INFO mapreduce.Job: Counters: 55
File System Counters
FILE: Number of bytes read=1007491124
FILE: Number of bytes written=1507738868
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=185540791
HDFS: Number of bytes written=478073828
HDFS: Number of read operations=11
HDFS: Number of large read operations=0
HDFS: Number of write operations=2
HDFS: Number of bytes read erasure-coded=0
Job Counters
Killed map tasks=1
Launched map tasks=3
Launched reduce tasks=1
Data-local map tasks=3
Total time spent by all maps in occupied slots (ms)=410756
Total time spent by all reduces in occupied slots (ms)=207107
Total time spent by all map tasks (ms)=410756
Total time spent by all reduce tasks (ms)=207107
Total vcore-milliseconds taken by all map tasks=410756
Total vcore-milliseconds taken by all reduce tasks=207107
Total megabyte-milliseconds taken by all map tasks=420614144
Total megabyte-milliseconds taken by all reduce tasks=212077568
Map-Reduce Framework
Map input records=3302326
Map output records=8866577
Map output bytes=563183844
Map output materialized bytes=499429681
Input split bytes=217
Combine input records=13805281
Combine output records=9584411
Reduce input groups=4645704
Reduce shuffle bytes=499429681
Reduce input records=4645707
Reduce output records=4645704
Spilled Records=14230118
Shuffled Maps =2
Failed Shuffles=0
Merged Map outputs=2
GC time elapsed (ms)=2649
CPU time spent (ms)=95210
Physical memory (bytes) snapshot=767778816
Virtual memory (bytes) snapshot=7587844096
Total committed heap usage (bytes)=608534528
Peak Map Physical memory (bytes)=392065024
Peak Map Virtual memory (bytes)=2529935360
Peak Reduce Physical memory (bytes)=164405248
Peak Reduce Virtual memory (bytes)=2533318656
Shuffle Errors
BAD_ID=0
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=185540574
File Output Format Counters
Bytes Written=478073828
#2.登錄歷史服務器地址
http://hadoop100:19888/jobhistory,查看到歷史任務列表

#3.查看任務運行日志

#4.運行日志詳情,至此Hadoop本地部署完成。
