个人服务器里面会装不少东西,因此我在跟目录下面创建了个doc文档文件夹java
1.建立存放软件的doc文件夹node
mkdir doc
2.进去doc文件夹进行下载hadoop-1.2.1资源包或者到个人百度云下载 地址http://pan.baidu.com/s/1gdSws07apache
cd doc wget http://mirror.bit.edu.cn/apache/hadoop/common/hadoop-1.2.1/hadoop-1.2.1.tar.gz
3.下载完毕进行解压hadoop-1.2.1.tar.gz服务器
tar -zxf hadoop-1.2.1.tar.gz
4.这个时候在doc目录应该有个hadoop-1.2.1文件夹,这个是咱们存放软件包的目录,通常咱们安装一个服务会单独的建立一个相关的文件夹,个人hadoop服务是装在app
/usr/local/hadoop/hadoop-1.2.1这个目录里面的,因此咱们把doc里面的hadoop-1.2.1复制到/usr/local/hadoop目录
#进入usr下面的local文件夹
cd /usr/local
#建立hadoop文件夹
mkdir hadoop
#转移hadoop-1.2.1文件夹到hadoop文件夹中
mv /doc/hadoop-1.2.1 /usr/local/hadoop
5.ok 如今开到配置hadoop的配置文件curl
配置hadoop-env.sh文件tcp
使用echo命令看JAVA_HOME,jdk的安装目录oop
[root@iZ94j7ahvuvZ conf]# echo $JAVA_HOME /usr/local/java/jdk1.7.0
修改hadoop-env.sh的JAVA_HOME信息fetch
进入hadoop的conf文件夹ui
cd /usr/local/hadoop/hadoop-1.2.1/conf
vi hadoop-env.sh
完善 JAVA_HOME属性
export JAVA_HOME=/usr/local/java/jdk1.7.0(本身的JDK目录)
6.配置文件core-site.xml
<configuration> <property> <name>hadoop.tmp.dir</name> <value>/hadoop</value> </property> <property> <name>dfs.name.dir</name> <value>/hadoop/name</value> </property> <property> <name>fs.default.name</name> <value>hdfs://localhost:9000</value> </property> </configuration>
<configuration> <property> <name>dfs.data.dir</name> <value>/hadoop/data</value> </property> </configuration>
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>localhost:9001</value>
</property>
</configuration>
cd /etc 在下面加入 export HADOOP_HOME=/usr/local/hadoop/hadoop-1.2.1 export PATH=$PATH:/usr/local/java/jdk1.7.0/bin:$HADOOP_HOME/bin(这里的JDK路径按照本身的JDK路径)
[root@iZ94j7ahvuvZ conf]# hadoop Usage: hadoop [--config confdir] COMMAND where COMMAND is one of: namenode -format format the DFS filesystem secondarynamenode run the DFS secondary namenode namenode run the DFS namenode datanode run a DFS datanode dfsadmin run a DFS admin client mradmin run a Map-Reduce admin client fsck run a DFS filesystem checking utility fs run a generic filesystem user client balancer run a cluster balancing utility oiv apply the offline fsimage viewer to an fsimage fetchdt fetch a delegation token from the NameNode jobtracker run the MapReduce job Tracker node pipes run a Pipes job tasktracker run a MapReduce task Tracker node historyserver run job history servers as a standalone daemon job manipulate MapReduce jobs queue get information regarding JobQueues version print the version jar <jar> run a jar file distcp <srcurl> <desturl> copy file or directories recursively distcp2 <srcurl> <desturl> DistCp version 2 archive -archiveName NAME -p <parent path> <src>* <dest> create a hadoop archive classpath prints the class path needed to get the Hadoop jar and the required libraries daemonlog get/set the log level for each daemon or CLASSNAME run the class named CLASSNAME
cd /usr/local/hadoop/hadoop-1.2.1/bin ./start-all.sh #敲完须要输入3次密码 以下提示则成功 [root@iZ94j7ahvuvZ bin]# ./start-all.sh namenode running as process 1341. Stop it first. root@localhost's password: localhost: starting datanode, logging to /usr/local/hadoop/hadoop-1.2.1/libexec/../logs/hadoop-root-datanode-iZ94j7ahvuvZ.out root@localhost's password: localhost: starting secondarynamenode, logging to /usr/local/hadoop/hadoop-1.2.1/libexec/../logs/hadoop-root-secondarynamenode-iZ94j7ahvuvZ.out starting jobtracker, logging to /usr/local/hadoop/hadoop-1.2.1/libexec/../logs/hadoop-root-jobtracker-iZ94j7ahvuvZ.out root@localhost's password: localhost: starting tasktracker, logging to /usr/local/hadoop/hadoop-1.2.1/libexec/../logs/hadoop-root-tasktracker-iZ94j7ahvuvZ.out