hue-3.9-cdh-5.7.0安装

前置依赖

[root@hadoop etc]# yum -y install ant asciidoc cyrus-sasl-devel cyrus-sasl-gssapi gcc gcc-c++ krb5-devel libtidy libxml2-devel libxslt-devel openldap-devel python-devel sqlite-devel openssl-devel mysql-devel gmp-devel

安装hue

将准备好的hue tar包解压;进入解压目录编译

[hadoop@hadoop hue-3.9.0-cdh5.7.0]$ make apps

进入apps目录下;看到编译完成的hue;配置环境变量
HUE_HOME=/home/hadoop/app/hue-3.9.0-cdh5.7.0
PATH=$HUE_HOME/build/env/lib:$PATH


进入desktop的conf目录下修改hue.ini配置文件和修改db权限【desktop目录下有配置文件hue.ini;build/env/lib下有启动命令】

[hadoop@hadoop hue-3.9.0-cdh5.7.0]$ cd desktop/conf/
[hadoop@hadoop conf]$ ll
总用量 56
-rw-r--r--. 1 hadoop hadoop 49070 3月  24 2016 hue.ini
-rw-r--r--. 1 hadoop hadoop  1843 3月  24 2016 log4j.properties
-rw-r--r--. 1 hadoop hadoop  1721 3月  24 2016 log.conf
[hadoop@hadoop conf]$ pwd
/home/hadoop/app/hue-3.9.0-cdh5.7.0/desktop/conf
[hadoop@hadoop conf]$ vim hue.ini

  # Set this to a random string, the longer the better.
  # This is used for secure hashing in the session store.
  secret_key=jFE93j;2[290-eiw.KEiwN2s3['d;/.q[eIW^y#e=+Iei*@Mn<qW5o

  # Webserver listens on this address and port
  http_host=hadoop
  http_port=8888

  # Time zone name
  time_zone=Asia/Shanghai

  配置hdfs的用户
  # This should be the hadoop cluster admin
  ## default_hdfs_superuser=hadoop





修改desktop的权限
[hadoop@hadoop hue-3.9.0-cdh5.7.0]$ cd desktop/
[hadoop@hadoop desktop]$ ll
总用量 252
drwxr-xr-x.  2 hadoop hadoop     81 4月  14 18:20 conf
drwxr-xr-x.  5 hadoop hadoop    183 4月  14 18:09 core
-rw-r--r--.  1 hadoop hadoop 253952 4月  14 18:14 desktop.db
drwxr-xr-x. 15 hadoop hadoop    210 3月  24 2016 libs
drwxrwxr-x.  2 hadoop hadoop     78 4月  14 18:13 logs
-rw-r--r--.  1 hadoop hadoop   3467 3月  24 2016 Makefile
[hadoop@hadoop desktop]$ chmod o+w desktop.db 
[hadoop@hadoop desktop]$ ll
总用量 252
drwxr-xr-x.  2 hadoop hadoop     81 4月  14 18:20 conf
drwxr-xr-x.  5 hadoop hadoop    183 4月  14 18:09 core
-rw-r--rw-.  1 hadoop hadoop 253952 4月  14 18:14 desktop.db
drwxr-xr-x. 15 hadoop hadoop    210 3月  24 2016 libs
drwxrwxr-x.  2 hadoop hadoop     78 4月  14 18:13 logs
-rw-r--r--.  1 hadoop hadoop   3467 3月  24 2016 Makefile


启动
[hadoop@hadoop bin]$ pwd
/home/hadoop/app/hue-3.9.0-cdh5.7.0/build/env/bin
[hadoop@hadoop bin]$ ./supervisor 

Hadoop集成环境相关添加

1、hdfs-site.xml

<property>
     <name>dfs.webhdfs.enabled</name>
     <value>true</value>
</property>

<property>
     <name>dfs.permissions.enabled</name>
     <value>false</value>
</property>

2、core-site.xml

<property>
     <name>hadoop.proxyuser.hue.hosts</name>
      <value>*</value>
</property>

<property>
      <name>hadoop.proxyuser.hue.groups</name>
      <value>*</value>
</property>

<property>
       <name>hadoop.proxyuser.hadoop.hosts</name>
       <value>*</value>
 </property>

  <property>
       <name>hadoop.proxyuser.hadoop.groups</name>
        <value>*</value>
 </property>

3、yarn-site.xml

<!--打开HDFS上日志记录功能-->
<property>
    <name>yarn.log-aggregation-enable</name>
    <value>true</value>
</property>

<!--在HDFS上聚合的日志最长保留多少秒。3天-->
<property>
     <name>yarn.log-aggregation.retain-seconds</name>
     <value>259200</value>
</property>

4、httpfs-site.xml

<property>
    <name>httpfs.proxyuser.hue.hosts</name>
    <value>*</value>
</property>

<property>
    <name>httpfs.proxyuser.hue.groups</name>
    <value>*</value>
</property>

5、配置文件同步

将以上配置文件同步到其他Hadoop主机

6、集成文件配置hive Hadoop MySQL

$HUE_HOME/desktop/conf/hue.ini

[hadoop]

  # Configuration for HDFS NameNode
  # ------------------------------------------------------------------------
  [[hdfs_clusters]]
    # HA support by using HttpFs

配置hdfs
    [[[default]]]
      # Enter the filesystem uri
      fs_defaultfs=hdfs://hadoop:8020

      # NameNode logical name.
      ## logical_name=

      # Use WebHdfs/HttpFs as the communication mechanism.
      # Domain should be the NameNode or HttpFs host.
      # Default port is 14000 for HttpFs.
      webhdfs_url=http://hadoop:50070/webhdfs/v1

      # Change this if your HDFS cluster is Kerberos-secured
      ## security_enabled=false

      # In secure mode (HTTPS), if SSL certificates from YARN Rest APIs
      # have to be verified against certificate authority
      ## ssl_cert_ca_verify=True

      # Directory of the Hadoop configuration
      ## hadoop_conf_dir=$HADOOP_CONF_DIR when set or '/etc/hadoop/conf'

  # Configuration for YARN (MR2)
  # ------------------------------------------------------------------------
[[yarn_clusters]]

配置yarn
    [[[default]]]
      # Enter the host on which you are running the ResourceManager
      resourcemanager_host=hadoop

      # The port where the ResourceManager IPC listens on
      resourcemanager_port=8032
	  
      #参考yarn-site.xml中的yarn.resourcemanager.address.rm1
      # Whether to submit jobs to this cluster
      submit_to=True

      # Resource Manager logical name (required for HA)
      ## logical_name=

      # Change this if your YARN cluster is Kerberos-secured
      ## security_enabled=false

      # URL of the ResourceManager API
      resourcemanager_api_url=http://hadoop:8088

      # URL of the ProxyServer API
      proxy_api_url=http://hadoop:8088

      # URL of the HistoryServer API
	  #参考mapred-site.xml中的mapreduce.jobhistory.webapp.address
      history_server_api_url=http://hadoop:19888

      # In secure mode (HTTPS), if SSL certificates from YARN Rest APIs
      # have to be verified against certificate authority
      ## ssl_cert_ca_verify=True

[beeswax]

配置hive
  # Host where HiveServer2 is running.
  # If Kerberos security is enabled, use fully-qualified domain name (FQDN).
  hive_server_host=hadoop

  # Port where HiveServer2 Thrift server runs on.
  hive_server_port=10000

配置MySQL
    # mysql, oracle, or postgresql configuration.
    ## [[[mysql]]]
      # Name to show in the UI.
       nice_name="My SQL DB"

      # For MySQL and PostgreSQL, name is the name of the database.
      # For Oracle, Name is instance of the Oracle server. For express edition
      # this is 'xe' by default.
       name=mysqldb

      # Database backend to use. This can be:
      # 1. mysql
      # 2. postgresql
      # 3. oracle
       engine=mysql

      # IP or hostname of the database to connect to.
       host=hadoop

      # Port the database server is listening to. Defaults are:
      # 1. MySQL: 3306
      # 2. PostgreSQL: 5432
      # 3. Oracle Express Edition: 1521
       port=3306

      # Username to authenticate with when connecting to the database.
       user=root

      # Password matching the username to authenticate with when
      # connecting to the database.
       password=root

7、Hive环境变量的添加

(hiveserver2,使用Mysql作为独立的元数据库)

hive-site.xml

<property>
    <name>hive.metastore.uris</name>
    <value>thrift://192.168.232.8:9083</value>
    <description>Thrift URI for the remote metastore. Used by metastore client to connect to remote metastore.</description>
  </property>

  <property>
    <name>hive.server2.thrift.bind.host</name>
    <value>192.168.232.8</value>
    <description>Bind host on which to run the HiveServer2 Thrift service.</description>
  </property>

启动hue

执行以下指令对hue数据库进行初始化

cd $HUE_HOME/build/env/

 bin/hue syncdb

 bin/hue migrate

启动顺序

1、启动Hive metastore
nohup hive --service metastore &

2、启动hiveserver2
nohup hive --service hiveserver2 &

3、启动Hue
nohup supervisor &

执行一个hive的查询:成功;同时可以查看hdfs文件目录;

879ded76ac2a6d5ac8105430ea28f3b86cd.jpg

f336a909089e4c968353d6283ca11240004.jpg

 

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值