以下是在CentOS上安装HDFS的核心步骤(基于Hadoop 3.x版本):
环境准备
sudo yum install java-1.8.0-openjdk-devel
echo "export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk" >> ~/.bashrc
source ~/.bashrc
/etc/hosts
)及SSH免密登录。下载与解压Hadoop
/usr/local
):wget https://archive.apache.org/dist/hadoop/common/hadoop-3.3.1/hadoop-3.3.1.tar.gz
tar -xzvf hadoop-3.3.1.tar.gz -C /usr/local/
sudo mv /usr/local/hadoop-3.3.1 /usr/local/hadoop
配置环境变量
/etc/profile.d/hadoop.sh
,添加:export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
使配置生效:source /etc/profile.d/hadoop.sh
。配置HDFS核心文件
core-site.xml
:设置默认文件系统(伪分布式可配置为localhost
):<property><name>fs.defaultFS</name><value>hdfs://namenode:9000</value></property>
hdfs-site.xml
:配置副本数、数据存储路径:<property><name>dfs.replication</name><value>3</value></property>
<property><name>dfs.namenode.name.dir</name><value>/usr/local/hadoop/data/namenode</value></property>
<property><name>dfs.datanode.data.dir</name><value>/usr/local/hadoop/data/datanode</value></property>
创建数据目录并授权:sudo mkdir -p /usr/local/hadoop/data/{namenode,datanode}
sudo chown -R hadoop:hadoop /usr/local/hadoop/data
格式化NameNode
hdfs namenode -format
启动HDFS集群
start-dfs.sh
(若需启动YARN,执行start-yarn.sh
)验证安装
jps
命令查看进程(需看到NameNode、DataNode等):jps
http://namenode:9870
)检查HDFS状态。注意事项:
参考来源: