在Ubuntu上启动HDFS前,需完成以下基础准备:
sudo apt update
sudo apt install openjdk-8-jdk
java -version # 验证安装(需显示Java版本信息)
/usr/local/)。wget https://downloads.apache.org/hadoop/common/hadoop-3.3.4/hadoop-3.3.4.tar.gz
tar -xzvf hadoop-3.3.4.tar.gz -C /usr/local/
~/.bashrc(或/etc/profile),添加Hadoop路径:export HADOOP_HOME=/usr/local/hadoop-3.3.4
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
source ~/.bashrc # 使配置生效
$HADOOP_HOME/etc/hadoop,编辑以下文件:
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/usr/local/hadoop-3.3.4/data/namenode</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/usr/local/hadoop-3.3.4/data/datanode</value>
</property>
</configuration>
sudo mkdir -p /usr/local/hadoop-3.3.4/data/namenode
sudo mkdir -p /usr/local/hadoop-3.3.4/data/datanode
sudo chown -R $USER:$USER /usr/local/hadoop-3.3.4/data # 当前用户拥有目录权限
hdfs namenode -format
start-dfs.sh脚本启动HDFS集群,该脚本会自动启动NameNode(主节点)和DataNode(数据节点)。start-dfs.sh
jps命令,若看到NameNode、DataNode进程,则说明启动成功。jps
http://localhost:9000(默认端口),若显示HDFS文件系统状态页面,则表示服务正常运行。sudo chown将Hadoop目录所有权赋予当前用户。9000端口被占用,可修改core-site.xml中的fs.defaultFS端口(如9001),并同步更新Web界面访问端口。core-site.xml、hdfs-site.xml的语法(如标签闭合),确保路径正确。