Debian环境下Hadoop更新升级步骤
/data/dfs/name、/data/dfs/data等目录)、YARN作业日志及Hadoop配置文件(core-site.xml、hdfs-site.xml、yarn-site.xml等),避免升级过程中因操作失误导致数据丢失。hadoop version命令)及Debian版本(通过lsb_release -a或cat /etc/debian_version),便于后续兼容性核查。sudo apt update && sudo apt upgrade -y && sudo apt full-upgrade -y
sudo apt autoremove -y && sudo apt clean
sudo systemctl stop hadoop-yarn-nodemanager hadoop-yarn-resourcemanager hadoop-mapreduce-historyserver
sudo systemctl stop hadoop-datanode hadoop-namenode
sudo apt install hadoop-<version> # 替换<version>为目标版本(如3.3.6)
/etc/profile.d/hadoop.sh)并添加至PATH。dfs.replication默认值变更、YARN的yarn.nodemanager.aux-services配置),重点检查路径、端口及新特性参数;cp -r /old/data/dfs/name/* /new/data/dfs/name/),并确保权限正确(chown -R hadoop:hadoop /new/data)。sudo systemctl start hadoop-namenode hadoop-datanode hadoop-yarn-resourcemanager hadoop-yarn-nodemanager hadoop-mapreduce-historyserver
hdfs namenode -rollingUpgrade prepare
hdfs namenode -rollingUpgrade start
# 依次重启各DataNode节点,完成滚动升级
hdfs namenode -rollingUpgrade finalize # 升级完成后确认
hadoop version;systemctl status hadoop-*,确保所有服务处于active (running)状态。hdfs dfsadmin -report(查看NameNode/Datanode状态)、hdfs dfs -put test.txt /tmp/(上传文件测试存储);yarn jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar pi 10 100(提交MapReduce作业,验证资源调度)。jps命令查看Hadoop进程是否正常启动;/var/log/hadoop-hdfs/、/var/log/hadoop-yarn/),确认无报错信息;http://<namenode-ip>:9870、ResourceManager的http://<resourcemanager-ip>:8088)监控集群状态。unattended-upgrades:sudo apt install unattended-upgrades
sudo dpkg-reconfigure --priority=low unattended-upgrades