我先在CentOS 7.4的服务器的/usr/
目录新建了一个hadoop目录,而后进入新建的hadoop目录,下载hadoop。html
命令:
wget http://mirrors.hust.edu.cn/apache/hadoop/common/hadoop-2.8.5/hadoop-2.8.5.tar.gz
node
到hadoop下载页面选择合适的hadoop版本进行下载,我使用的是2.8.5,截止如今,由于hbase最高支持2.8.X的版本,使用Hadoop2.9.X会有问题。linux
把下载好的压缩包解压出来apache
命令:
tar xvzf hadoop-2.8.5.tar.gz
vim
配置/etc/profile的JAVA_HOME服务器
配置${HADOOP_HOME}/etc/hadoop/core-site.xml
less
<configuration> <property> <name>fs.defaultFS</name> <value>hdfs://localhost:9000</value> </property> </configuration>
配置${HADOOP_HOME}/etc/hadoop/hdfs-site.xml
这个配置是配置HDFS的备份数的,不配置的话默认是备份3分,由于这里是测试,因此1份就够了。通常用默认配置,土豪能够调3+。ssh
<configuration> <property> <name>dfs.replication</name> <value>1</value> </property> </configuration>
如今经过ssh localhost
看看需不须要密码访问。若是发现须要密码访问,那么试用一下命令配置ssh-keygen
:分布式
ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys chmod 0600 ~/.ssh/authorized_keys
${HADOOP_HOME}/bin/hdfs namenode -format
${HADOOP_HOME}/sbin/start-dfs.sh
${HADOOP_HOME}/bin/hdfs dfs -mkdir /user ${HADOOP_HOME}/bin/hdfs dfs -mkdir /user/<username>
${HADOOP_HOME}/bin/hdfs dfs -put etc/hadoop input
${HADOOP_HOME}/bin/hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.8.5.jar grep input output 'dfs[a-z.]+'
跑完会输出这个:工具
最后中止运行:
${HADOOP_HOME}/sbin/stop-dfs.sh
先配置MapReduce的管理工具为yarn:
cp ${HADOOP_HOME}/etc/hadoop/mapred-site.xml.template ${HADOOP_HOME}/etc/hadoop/mapred-site.xml vim ${HADOOP_HOME}/etc/hadoop/mapred-site.xml
mapred-site.xml配置以下:
<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> </configuration>
配置yarn-site.xml:
$ vim ${HADOOP_HOME}/etc/hadoop/yarn-site.xml <configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> </configuration>
启动yarn:
${HADOOP_HOME}/sbin/start-yarn.sh jps 32934 NodeManager 32999 Jps 32857 ResourceManager