以下是在Linux上配置Hadoop的步骤:
sudo apt update和sudo apt install openjdk-11-jdk,安装完成后用java -version验证。wget https://downloads.apache.org/hadoop/common/hadoop-3.3.1/hadoop-3.3.1.tar.gz,然后解压到指定目录,如/usr/local,使用命令tar -xzvf hadoop-3.3.1.tar.gz和sudo mv hadoop-3.3.1 /usr/local/hadoop。~/.bashrc文件,添加export HADOOP_HOME=/usr/local/hadoop、export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin等,执行source ~/.bashrc使配置生效。$HADOOP_HOME/etc/hadoop目录下,修改core-site.xml、hdfs-site.xml等文件。例如,在core-site.xml中设置fs.defaultFS为hdfs://localhost:9000,在hdfs-site.xml中设置dfs.replication为1。ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa,将公钥复制到其他节点cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys。hdfs namenode -format。start-dfs.sh和start-yarn.sh命令启动。jps命令查看进程,或通过浏览器访问http://localhost:9870(HDFS Web界面)和http://localhost:8088(YARN ResourceManager)验证。