温馨提示×

温馨提示×

您好,登录后才能下订单哦!

密码登录×
登录注册×
其他方式登录
点击 登录注册 即表示同意《亿速云用户服务条款》

安装hadoop

发布时间:2020-07-08 06:45:29 来源:网络 阅读:531 作者:liufu1103 栏目:大数据

1:安装包

http://hadoop.apache.org/releases.html

http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html

hadoop-2.6.4 (binary)

jdk-8u102-linux-x64.tar.gz


2:节点信息

centos7 * 3

master 10.10.0.115

slave1 10.10.0.116

slave2 10.10.0.117


3:安装过程

3.1 节下初始化(略)

     主机名  防火墙  selinux  等

3.2 节点互信

[root@master ~]#ssh-keygen

[root@master ~]#cat /root/.ssh/id_rsa.pub >  /root/.ssh/authorized_keys 

[root@master ~]#scp -r /root/.ssh slave1:/root/

[root@master ~]#scp -r /root/.ssh slave2:/root/


3.3 在所有节点安装jdk

[root@master ~]#tar -zxvf jdk-8u102-linux-x64.tar.gz

[root@master ~]#mkdir   /usr/soft

[root@master ~]#mv jdk1.8.0_102 /usr/soft/

[root@master ~]#vim /etc/profile

   ...

   export JAVA_HOME=/usr/soft/jdk1.8.0_102
   export CLASSPATH=.:$JAVA_HOME/jre/lib/rt.jar:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
   export PATH=$PATH:$JAVA_HOME/bin


[root@master ~]#source /etc/profile

[root@master ~]#java -version
java version "1.8.0_102"
Java(TM) SE Runtime Environment (build 1.8.0_102-b14)
Java HotSpot(TM) 64-Bit Server VM (build 25.102-b14, mixed mode)


其它节点同样操作


3.4 安装hadoop

先在master上安装配置,然后scp到其它节点即可

[root@master ~]#tar -zxvf hadoop-2.6.4.tar.gz

[root@master ~]#mv hadoop-2.6.4 /usr/soft/

[root@master ~]#cd /usr/soft/hadoop-2.6.4/

[root@master ~]#mkdir hdfs

[root@master ~]#mkdir hdfs/data

[root@master ~]#mkdir hdfs/name

[root@master ~]#mkdir tmp

[root@master ~]# cat /usr/soft/hadoop-2.6.4/etc/hadoop/core-site.xml

<configuration>
   <property>
        <name>fs.defaultFS</name>
        <value>hdfs://10.10.0.115:9000</value>
    </property>
    <property>
        <name>hadoop.tmp.dir</name>
        <value>file:/usr/soft/hadoop-2.6.4/tmp</value>
    </property>
    <property>
        <name>io.file.buffer.size</name>
        <value>131702</value>
    </property>
</configuration>


[root@master ~]# cat /usr/soft/hadoop-2.6.4/etc/hadoop/hdfs-site.xml

<configuration>
    <property>
        <name>dfs.namenode.name.dir</name>
        <value>file:/usr/soft/hadoop-2.6.4/hdfs/name</value>
    </property>
    <property>
        <name>dfs.datanode.data.dir</name>
        <value>file:/usr/soft/hadoop-2.6.4/hdfs/data</value>
    </property>
    <property>
        <name>dfs.replication</name>
        <value>2</value>
    </property>
    <property>
        <name>dfs.namenode.secondary.http-address</name>
        <value>10.10.0.115:9001</value>
    </property>
    <property>
    <name>dfs.webhdfs.enabled</name>
    <value>true</value>
    </property>
</configuration>


 [root@master ~]#cp /usr/soft/hadoop-2.6.4/etc/hadoop/mapred-site.xml.template   /usr/soft/hadoop-2.6.4/etc/hadoop/mapred-site.xml


[root@master ~]# cat /usr/soft/hadoop-2.6.4/etc/hadoop/mapred-site.xml

<configuration>
   <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
    </property>
    <property>
        <name>mapreduce.jobhistory.address</name>
        <value>10.10.0.115:10020</value>
    </property>
    <property>
        <name>mapreduce.jobhistory.webapp.address</name>
        <value>10.10.0.115:19888</value>
    </property>
    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
    <property>
        <name>yarn.nodemanager.auxservices.mapreduce.shuffle.class</name>
        <value>org.apache.hadoop.mapred.ShuffleHandler</value>
    </property>
    <property>
        <name>yarn.resourcemanager.address</name>
        <value>10.10.0.115:8032</value>
    </property>
    <property>
        <name>yarn.resourcemanager.scheduler.address</name>
        <value>10.10.0.115:8030</value>
    </property>
    <property>
        <name>yarn.resourcemanager.resource-tracker.address</name>
        <value>10.10.0.115:8031</value>
    </property>
    <property>
        <name>yarn.resourcemanager.admin.address</name>
        <value>10.10.0.115:8033</value>
    </property>
    <property>
        <name>yarn.resourcemanager.webapp.address</name>
        <value>10.10.0.115:8088</value>
    </property>
    <property>
        <name>yarn.nodemanager.resource.memory-mb</name>
        <value>768</value>
    </property>
</configuration>


[root@master ~]# cat /usr/soft/hadoop-2.6.4/etc/hadoop/hadoop-env.sh

...

export JAVA_HOME=/usr/soft/jdk1.8.0_102


[root@master ~]# cat /usr/soft/hadoop-2.6.4/etc/hadoop/yarn-env.sh

...

export JAVA_HOME=/usr/soft/jdk1.8.0_102


[root@master ~]# cat  /usr/soft/hadoop-2.6.4/etc/hadoop/slaves
10.10.0.116
10.10.0.117


[root@master ~]#scp -r  /usr/soft/jdk1.8.0_102 slave1:/usr/soft/

[root@master ~]#scp -r  /usr/soft/jdk1.8.0_102 slave2:/usr/soft/


[root@master ~]#/usr/soft/hadoop-2.6.4/bin/hdfs namenode -format

[root@master ~]# /usr/soft/hadoop-2.6.4/sbin/start-all.sh



浏览器打开http://10.10.0.115:8088/
浏览器打开http://10.10.0.115:50070/








  





向AI问一下细节

免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。

AI