1.节点说明
| ip |
hostname |
说明 |
| 10.0.0.120 |
hadoop-all |
主节点+数据节点 |
2.环境准备
需要JDK1.8运行环境,搭建请参考[Jdk安装(Linux,MacOS,Windows)]
3.设置主机名
hostnamectl set-hostname hadoop-all
4.设置host
echo "10.0.0.120 hadoop-all" >> /etc/hosts
5.设置ssh免密码登录(主节点)
ssh-keygen -t rsa
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
ssh hadoop-all
6.开放端口
firewall-cmd --add-port=9870/tcp && firewall-cmd --add-port=9870/tcp --permanent
firewall-cmd --add-port=8088/tcp && firewall-cmd --add-port=8088/tcp --permanent
firewall-cmd --add-port=9000/tcp && firewall-cmd --add-port=9000/tcp --permanent
firewall-cmd --add-port=50090/tcp && firewall-cmd --add-port=50090/tcp --permanent
systemctl stop firewalld.service && systemctl disable firewalld.service
7.准备hadoop安装包
http://hadoop.apache.org/
https://mirrors.cloud.tencent.com/apache/hadoop/common/
mkdir -p /data0/hadoop/
tar -zxvf hadoop-3.3.6.tar.gz
mkdir -p /data1/hadoop/repo
8.修改配置
8.1.进入配置文件目录
cd /data0/hadoop/hadoop-3.3.6/etc/hadoop
8.2.修改hadoop-env.sh配置文件
vi hadoop-env.sh
export JAVA_HOME=/data0/java/jdk1.8.0_361
export HADOOP_LOG_DIR=/data1/hadoop/repo/logs/hadoop
8.3.修改core-site.xml配置文件
vi core-site.xml
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://hadoop-all:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/data1/hadoop/repo</value>
</property>
</configuration>
8.4.修改hdfs-site.xml配置文件
vi hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
8.5.修改mapred-site.xml配置文件
vi mapred-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
8.6.修改yarn-site.xml配置文件
vi yarn-site.xml
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.env-whitelist</name>
<value>JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME,HADOOP_MAPRED_HOME</value>
</property>
</configuration>
9.修改hadoop启动/停止配置
9.1.进入启动脚本目录
cd /data0/hadoop/hadoop-3.3.6/sbin
9.2.修改dfs.sh的启动/停止命令
vi start-dfs.sh
vi stop-dfs.sh
HDFS_DATANODE_USER=root
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root
HDFS_SECONDARYNAMENODE_USER=root
9.3.修改yarn.sh的启动/停止命令
vi start-yarn.sh
vi stop-yarn.sh
YARN_RESOURCEMANAGER_USER=root
HADOOP_SECURE_DN_USER=yarn
YARN_NODEMANAGER_USER=root
10.格式化hdfs
cd /data0/hadoop/hadoop-3.3.6
bin/hdfs namenode -format
11.启动/停止
cd /data0/hadoop/hadoop-3.3.6/sbin
./start-all.sh
./stop-all.sh
jps
12.web查看状态
10.0.0.120:9870
10.0.0.10:8088
13.配置环境变量
vi ~/.bash_profile
export HADOOP_HOME=/data0/hadoop/hadoop-3.3.6
export PATH=$HADOOP_HOME/bin:$PATH
:x
source ~/.bash_profile