Debian环境下Hadoop升级步骤
/user目录下的文件)、NameNode元数据(/var/lib/hadoop-hdfs/cache/hadoop-hdfs/dfs/name/current目录)及所有配置文件(core-site.xml、hdfs-site.xml、yarn-site.xml等),防止升级过程中数据丢失。hadoop version确认当前Hadoop版本;检查Debian系统版本(lsb_release -a或cat /etc/debian_version),确保目标Hadoop版本与Debian版本兼容(如Hadoop 3.x需搭配Debian 10及以上版本)。sudo apt updatesudo apt upgradesudo apt full-upgradesudo apt autoremove + sudo apt clean(释放磁盘空间)。hadoop-3.3.6.tar.gz),选择与Debian系统架构(amd64/arm64)匹配的版本。/opt/hadoop),设置目录权限(chown -R hadoop:hadoop /opt/hadoop)。/opt/hadoop-old),便于后续回滚。core-site.xml、hdfs-site.xml、yarn-site.xml、mapred-site.xml等复制到新版本对应目录,根据新版本要求修改配置(如Hadoop 3.x需调整yarn.nodemanager.aux-services为mapreduce_shuffle)。JAVA_HOME环境变量(/etc/profile或~/.bashrc)。sudo systemctl stop hadoop-namenodesudo systemctl stop hadoop-datanodesudo systemctl stop hadoop-yarn-resourcemanagersudo systemctl stop hadoop-yarn-nodemanagersudo systemctl stop hadoop-jobhistoryserver。/opt/hadoop)覆盖旧版本目录。hadoop namenode -upgrade(根据提示完成元数据升级)。/tmp/hadoop-*)。sudo systemctl start hadoop-namenode → sudo systemctl start hadoop-datanodesudo systemctl start hadoop-yarn-resourcemanager → sudo systemctl start hadoop-yarn-nodemanagersudo systemctl start hadoop-jobhistoryserver。hadoop version,确认显示为目标版本。jps查看进程(应包含NameNode、DataNode、ResourceManager等);通过hdfs dfsadmin -report查看HDFS节点状态;通过yarn node -list查看YARN节点列表。hdfs dfs -put/hdfs dfs -get)和MapReduce作业(如hadoop jar hadoop-mapreduce-examples.jar pi 10 100),验证功能正常。unattended-upgrades包(sudo apt install unattended-upgrades),配置自动安装安全更新,保持系统安全。tail -f /var/log/hadoop-hdfs/*.log监控日志,调整配置参数(如dfs.replication、mapreduce.map.memory.mb)优化性能;使用hadoop balancer.sh平衡数据分布。