hadoop-2.10.0安装hive-2.3.6

公司创建数仓,hive是必不可少的,hive是创建在hadoop基础上的数据库,前面已经搭建起了hadoop高可用,要学习hive,先从搭建开始,下面梳理一下hive搭建过程html

1.下载hive安装包 ,下载地址:https://hive.apache.org/downloads.html java

找到本身hadoop对应的版本下载安装,我这里下载的是 apache-hive-2.3.6-bin.tar.gzmysql

2.安装hive,将安装包解压到/opt/soft下,并创建软连接sql

tar -zxvf apache-hive-2.3.6-bin.tar.gz -C /opt/soft/
cd /opt/soft
mv apache-hive-2.3.6-bin hive-2.3.6
ln -s hive-2.3.6 hive

3.配置环境变量/etc/profile数据库

vim /etc/profile

#hive export HIVE_HOME
=/opt/soft/hive export PATH=$PATH:$HIVE_HOME/bin

保存后别忘记编译一下apache

source /etc/profile

4.配置hive配置文件,hive元数据默认存储到derby数据库中,咱们这里使用mysql来存储,hive-site.xml配置信息较多建议下载到windows下修改,而后再传上去vim

首先复制默认的配置文件模板,里面已经包含hive全部的默认配置信息windows

cp hive-default.xml.template hive-site.xml

修改hive-site.xml配置文件,将元数据存放数据库改成mysql,在hive-site.xml中找到下列属性,修改成:centos

<property>
    <name>javax.jdo.option.ConnectionDriverName</name>
    <value>com.mysql.jdbc.Driver</value>
    <description>Driver class name for a JDBC metastore</description>
  </property>
  <property>
    <name>javax.jdo.option.ConnectionURL</name>
    <value>jdbc:mysql://192.168.118.1:3306/hive2</value>
    <description>
      JDBC connect string for a JDBC metastore.
      To use SSL to encrypt/authenticate the connection, provide database-specific SSL flag in the connection URL.
      For example, jdbc:postgresql://myhost/db?ssl=true for postgres database.
    </description>
  </property>
  <property>
    <name>javax.jdo.option.ConnectionUserName</name>
    <value>root</value>
    <description>Username to use against metastore database</description>
  </property>
  <property>
    <name>javax.jdo.option.ConnectionPassword</name>
    <value>111111</value>
    <description>password to use against metastore database</description>
  </property>

数据库驱动为mysql驱动com.mysql.jdbc.Driver,URL改成mysql的hive2(自定义)数据库,用户名密码为本身数据库对应的用户名密码ide

修改hive配置的一些目录,指定到本身选择的目录,搜索以 ${system 开头的 value 替换为本身的目录,我这里替换为:/home/hdfs/hive下相关目录

  <property>
    <name>hive.exec.local.scratchdir</name>
    <value>/home/hdfs/hive</value>
    <description>Local scratch space for Hive jobs</description>
  </property>
<property>
    <name>hive.downloaded.resources.dir</name>
    <value>/home/hdfs/hive/downloads</value>
    <description>Temporary local directory for added resources in the remote file system.</description>
  </property>
  <property>
    <name>hive.querylog.location</name>
    <value>/home/hdfs/hive/querylog</value>
    <description>Location of Hive run time structured log file</description>
  </property>
  <property>
    <name>hive.server2.logging.operation.log.location</name>
    <value>/home/hdfs/hive/server2_logs</value>
    <description>Top level directory where operation logs are stored if logging functionality is enabled</description>
  </property>

修改权限验证为false

 <property>
    <name>hive.server2.enable.doAs</name>
    <value>false</value>
    <description>
      Setting this property to true will have HiveServer2 execute
      Hive operations as the user making the calls to it.
    </description>
  </property>

5.既然修改元数据存放在mysql库里,就须要将mysql驱动包放入到hive/lib中,注意mysql版本和驱动包一致

mv mysql-connector-java-8.0.18.jar /opt/soft/hive/lib/

6.在mysql数据库中建立hive2库

7.初始化hive的元数据(表结构)到mysql中。

cd /opt/soft/hive/bin
schematool -dbType mysql -initSchema

出现以下信息,表明成功

 

也能够查看mysql中hive2库,全部表初始化完成

 

 

8.启动hadoop,若是没有安装能够参考:centos7搭建hadoop2.10高可用(HA)

start-all.sh

9.启动hive

hive

 

 查看目前只有default数据库

建立数据库:

create database myhivedb2;

 

 查看建立的mysqhivedb2已经出来了

咱们查一下hdfs中是否建立了对应的目录

hdfs dfs -ls -R /user/hive/

 

 也能够查看mysql中hive2库的dbs表:

 

 至此hive环境搭建完成

相关文章
相关标签/搜索