java操作hive获取数据上传hdfs并maven打包依赖本地jar 带源码

背景

大数据平台配置可执行的jar包,需求是jar包能够获取hive表数据,并将数据上传到hdfs。

组件

jdk8 + hive + hdfs

源码

https://gitee.com/acelee723/acelee-hive-hdfs-main-jar

代码 

1.hive操作类

import org.mortbay.util.ajax.JSON;

import java.sql.*;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;

/**
 * Hive操作类
 *
 * @author Ace Lee
 * @date 2019/8/27 14:18
 * @version 1.0
 **/
public class HiveClientUtils {

    private static String driverName = "com.cloudera.hive.jdbc4.HS2Driver";

    //填写hive的IP,之前在配置文件中配置的IP
    private static String Url = "jdbc:hive2://10.10.10.10:10000/hbzfw";

    private static Connection conn;

    private static PreparedStatement ps;

    private static ResultSet rs;

    //创建连接
    public static Connection getConnnection() {
        try {
            Class.forName(driverName);
            //此处的用户名一定是有权限操作HDFS的用户,否则程序会提示"permission deny"异常
            conn = DriverManager.getConnection(Url, "", "");
        } catch (ClassNotFoundException e) {
            e.printStackTrace();
            System.exit(1);
        } catch (SQLException e) {
            e.printStackTrace();
        }
        return conn;
    }

    public static PreparedStatement prepare(Connection conn, String sql) {
        PreparedStatement ps = null;
        try {
            ps = conn.prepareStatement(sql);
        } catch (SQLException e) {
            e.printStackTrace();
        }
        return ps;
    }

    public static List<Map<String,Object>> getResult(String sql) {
        List<Map<String,Object>> rowDatas = new ArrayList<Map<String, Object>>();
        System.out.println(sql);
        conn = getConnnection();

        try {
            ps = prepare(conn, sql);
            rs = ps.executeQuery();
            ResultSetMetaData md = rs.getMetaData();
            int columCount = md.getColumnCount();

            while(rs.next()) {
                Map<String,Object> rowData = new HashMap<String, Object>();
                for(int i = 1; i <= columCount; i++) {
                    rowData.put(md.getColumnName(i), rs.getObject(i));
                }
                rowDatas.add(rowData);
            }
            System.out.println(JSON.toString(rowDatas));
        } catch (SQLException e) {
            e.printStackTrace();
        }
        return rowDatas;
    }

}

2.hdfs操作类

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IOUtils;

import java.io.*;
import java.net.URI;
import java.text.SimpleDateFormat;
import java.util.Date;

/**
 * Hdfs操作类
 *
 * @author Ace Lee
 * @date 2019/8/27 14:18
 * @version 1.0
 **/
public class HdfsFileSystem {

    /**
     * HDFS集群路径
     */
    private static final String hdfsPath = "hdfs://10.10.10.10:8020";

    public static void copyFileToHDFSByName(Configuration conf,String localFileName, String remoteFileName) throws IOException {
        FileSystem fs = FileSystem.get(URI.create(hdfsPath), conf);
        fs.copyFromLocalFile(new Path(localFileName), new Path(remoteFileName));
        System.out.println("copy from local file:" + localFileName + " to HDFS file:" + remoteFileName + " done.");
        fs.close();
    }

    public static void copyFileToHDFSByFileObj(byte[] bytes, String prefix, String fileName) throws IOException {
        InputStream in =
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值