安装目标:一个namenode 和 2 个datanode,假定已用户名zhaomeng和密码zhaomeng进行java
虚拟机:3个centos 64(桥接模式)node
1.192.168.203.129 master_centos64_1 2.192.168.203.130 slave_centos64_2 3.192.168.203.138 slave_centos_3
准备工做:web
1.启动三个虚拟机master_centos64_1centos
2.检查ssh是否安装和sshd是否启动(能否互相ssh,端口号默认是22)ssh
3.若想配置好记的机器名,能够sudo vi /etc/hostname将HOSTNAME改为本身喜欢的oop
4.若想只用机器名,能够sudo vi /etc/hosts更改路由表 (hosts文件优先于路由器)测试
5.(optional)检查防火墙iptablecode
6.在三台机器上建立相同用户,如zhaomeng, 加入sudoers列表和root用户组orm
a.useradd -G root zhaomengxml
b.sudovi /etc/sudoers 正加root相似的一行
root ALL(ALL) ALL zhaomeng ALL(ALL) ALL
7.配置用户zhaomeng在三台机器相互间免密码ssh
a.(master1)cd /home/zhaomeng
b.(master1)ssh-keygen以后直接3个回车或者ssh-keygen -t rsa -P '' (在主目录生成.ssh/idrsa和.ssh、id_rsa.pub)
c.(master1)cat .ssh/id_rsa.pub >> .ssh/authorized_keys
d.(slave2和slave3)执行ab并将id_rsa.pub文件均scp至master1,并执行c
e.(master1)将包含3个公钥的authorized_keys文件scp至slave2和slave3
f.(master1,slave2,slave3)chmod 600 .ssh/authorized_keys(很是重要,不然不起效)
g.(master1,slave2,slave3)三台机器之间相互测试时候能够无密码ssh (第一次均须要确认是否记住凭证,选择yes)
8.(master1,slave2,slave3)下载和安装JDK7,并设置必要的JAVA_HOME和PATH变量
9.(master1)下载hadoop2.6.3,解压至/home/zhaomeng/hadoop263 (hadoop263下路目录为bin sbin etc libexec lib 等)
10.(master1,slave2,slave3)设置环境变量
vi /etc/profile 加入以下内容
export JAVA_HOME=/usr/java/default export HADOOP_HOME=/home/zhaomeng/hadoop export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop export HADOOP_PREFIX=$HADOOP_HOME export HADOOP_LIBEXEC_DIR=$HADOOP_HOME/libexec export PATH=$JAVA_HOME/bin:$PATH export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native export HADOOP_COMMON_HOME=$HADOOP_HOME export HADOOP_HDFS_HOME=$HADOOP_HOME export HADOOP_MAPRED_HOME=$HADOOP_HOME export HADOOP_YARN_HOME=$HADOOP_HOME
并使之生效source /etc/profile
11.(master1)配置HADOOP_HOME/etc/hadoop/core.site.xml hdfs.site.xml yarn.site.xml (mapred-site.xml) slaves 以及hadoop-sne.sh和yarn-env.sh (两个sh文件主要是加入JAVA_HOME变量)
12.(master1)mkdir logs tmp data
13.(master1)将HADOOP_HOME整个文件夹scp至slave2和slave3,并保持和master1同样的目录结构
14.(master1)格式化namenode,执行HADOOP_HOME/bin/hdfs namenode -format
15.(master1)启动hdfs,执行HADOOP_HOME/sbin/start-dfs.sh
16.(master1)启动yarn,执行HADOOP_HOME/sbin/start-yarn.sh
17.几个网址
18.待续
附录几个文件
core-site.xml
<configuration> <property> <name>fs.defaultFS</name> <value>hdfs://192.168.203.129:9000</value> </property> <property> <name>hadoop.tmp.dir</name> <value>/home/zhaomeng/hadoop/tmp</value> </property> </configuration>
hdfs-site.xml
<configuration> <property> <name>dfs.webhdfs.enable</name> <value>true</value> </property> <property> <name>dfs.datanode.ipc.address </name> <value>0.0.0.0:50020</value> </property> <property> <name>dfs.datanode.http.address</name> <value>0.0.0.0:50075</value> </property> <property> <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:///home/zhaomeng/hadoop/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:///home/zhaomeng/hadoop/data</value> </property> </configuration>
yarn-site.xml
<configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> </configuration>
slaves
slave_centos64_2 slave_centos_3