export JAVA_HOME=/usr/local/jdk1.8.0_73 export JRE_HOME=${JAVA_HOME}/jre export CLASSPATH=.:${JAVA_HOME}/lib:${JRE_HOME}/lib export PATH=$PATH:${JAVA_HOME}/bin
不要忘了source /etc/profile哦!
b).建立hadoop用户html
# 建立hadoop用户 useradd -m hadoop #设置hadoop用户密码 passwd hadoop
export HADOOP_HOME=/home/hadoop/hadoop-2.7.2 export PATH=$PATH:${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin
b).进入hadoop_home/etc/hadoop/目录,修改以下文件内容
1.修改hadoop-env.sh文件,配置JAVA_HOMEnode
export JAVA_HOME=/usr/local/jdk1.8.0_73
2.修改core-site.xml文件添加以下内容web
<configuration> <property> <name>fs.defaultFS</name> <value>hdfs://hadoop-master:9000</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/home/hadoop/hadoop-2.7.2/tmp</value> </property> </configuration>
3.修改hdfs-site.xml文件,添加以下内容apache
<configuration> <property> <name>dfs.namenode.name.dir</name> <value>file:/home/hadoop/hadoop-2.7.2/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/home/hadoop/hadoop-2.7.2/dfs/data</value> </property> <property> <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.namenode.secondary.http-address</name> <value>hadoop-master:9001</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property> </configuration>
4.修改mapred-site.xml文件,添加以下内容ubuntu
<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>hadoop-master:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>hadoop-master:19888</value> </property> </configuration>
5.修改yarn-site.xml文件,添加以下内容bash
<configuration> <property> <name>yarn.resourcemanager.address</name> <value>hadoop-master:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>hadoop-master:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>hadoop-master:8031</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>hadoop-master:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>hadoop-master:8088</value> </property> <property> <name>yarn.resourcemanager.hostname</name> <value>hadoop-master</value> </property> <property> <name>yarn.resourcemanager.scheduler.class</name> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value> </property> <property> <name>yarn.scheduler.maximum-allocation-mb</name> <value>1536</value> </property> <!-- config node manager --> <property> <name>yarn.nodemanager.resource.memory-mb</name> <value>1536</value> </property> <property> <name>yarn.nodemanager.resource.cpu-vcores</name> <value>8</value> </property> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> </configuration>
6.修改slaves文件,将全部从节点添加到该文件中,如:服务器
#这里的节点名次须要和/etc/hosts文件中配置映射的名称一致 hadoop-slave1 hadoop-slave2
修改主机名为hadoop-master(/etc/hostname)
注意:每个从节点都要修改相应的hostname,如hadoop-slave1,hadoop-slave2,.....oracle
在主节点上格式化hdfs,hadoop namenode -formatapp
修改hosts文件(每个主从节点都要配置),将主从节点都配置进去,注释原有的localhost,以下:ssh
#127.0.0.1 localhost #127.0.1.1 ubuntu #主节点 192.168.100.180 hadoop-master #从节点1 192.168.100.181 hadoop-slave1 #从节点2 192.168.100.182 hadoop-slave2
将配置ssh免密码登,这里不赘述,参考:http://jingyan.baidu.com/article/f3ad7d0fe7bd0d09c2345b75.html
克隆当前配置的服务器系统到多个本身点或在多个从节点上重复1~5步骤。
注意:1.克隆系统后记得修改/etc/hostname文件,/etc/hosts文件主从节点保持一致。
2.克隆后须要修改子节点的ip地址,建议将主从节点的ip设置为静态IP。
经过以上配置后,基本完成hadoop的环境配置工做,可经过start-all.sh启动hadoop
a).经过master的主机IP:8088访问hadoop的web管理界面,如图
b).经过master的主机IP:50070访问hdfs的web管理界面