hadoop伪分布式集群的安装(不是单机版)
准备工作三台虚拟机,关闭防火墙,关闭selinux
查看防火状态 systemctl status firewalld
暂时关闭防火墙 systemctl stop firewalld
永久关闭防火墙 systemctl disable firewalld
查看 selinux状态 getenforce
暂时关闭 selinux setenforce 0
永久关闭 selinux 在/etc/selinux/config文件中将SELINUX改为disabled
修改主机名称
三台主机
hostnamectl set-hostname master
hostnamectl set-hostname slave1
hostnamectl set-hostname slave2<br><br>使用bash命令刷新生效在/etc/hosts文件中添加ip映射
IP+主机名称根据自己需求修改,这里给出模板
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152102045-1071353637.png
配置ssh免密登录
ssh-keygten -t rsa #生成密钥
ssh-copy-id master#分发给其他节点,分发给自己主要是为了之后群集集群不需要输入密码
ssh-copy-id slave1
ssh-copy-id slave2安装JAVA和HADOOP
解压JAVA
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152342030-1853758970.png
解压HADOOP
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152403538-731260551.png
修改名称为jdk与hadoop
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152444146-1428694292.png
配置环境变量
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152507321-965357235.png
刷新环境变量,使生效
source /etc/profile使用javac 与hadoop verison验证是否安装成功
配置hadoop文件
core-site.xml文件
fs.defaultFS
hdfs://hadoop:9000
hadoop.tmp.dir
/usr/local/src/hadoop-2.7.2/data/tmp
hdfs-site.xml
dfs.replication
1
dfs.namenode.name.dir
file:/home/hadoop/hadoop-2.6.0/hdfs/name
true
dfs.datanode.data.dir
file:/home/hadoop/hadoop-2.6.0/hdfs/data
true
dfs.namenode.secondary.http-address
master:9001
dfs.webhdfs.enabled
true
dfs.permissions
false
yarn.site.xml
yarn.resourcemanager.address
master:18040
yarn.resourcemanager.scheduler.address
master:18030
yarn.resourcemanager.webapp.address
master:18088
yarn.resourcemanager.resource-tracker.address
master:18025
yarn.resourcemanager.admin.address
master:18141
yarn.nodemanager.aux-services
mapreduce_shuffle
yarn.nodemanager.auxservices.mapreduce.shuffle.class
org.apache.hadoop.mapred.ShuffleHandler
mapred-site.xml
mapreduce.framework.name
yarn
在hadoop-env.sh yarn-env.sh mapred-env.sh中配置java环境
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152920249-1240928536.png
这里给出hadoop-env.sh 其他相同
配置slave
写入三台主机的主机名
https://img2023.cnblogs.com/blog/3085047/202304/3085047-20230414152959684-1375104890.png
将配置好的hadoop分发给其他主机
# scp -r /usr/local/src/hadoop/ root@slave1:/usr/local/src/
# scp -r /usr/local/src/hadoop/ root@slave2:/usr/local/src/格式化namenode
hdfs namenode -format启动dfs
start-dfs.sh启动yarn
start-yarn.sh使用jps查看
master节点:ResourceManager,DataNode,SecondaryNameNode,NameNode,NodeManager
slave1节点与slave2节点:NodeManager,DataNode
集群全部启动则为启动成功
进入web页面验证
namenode web页面 = IP+50070
yarn web页面 =IP 8088
免责声明:如果侵犯了您的权益,请联系站长,我们会及时删除侵权内容,谢谢合作!
页:
[1]