<返回更多

JDBC连接池封装MaxCompute/Hive/Oracle/Mysql

2022-08-17    IT动力
加入收藏

有时候需要连接第三方的各种数据源,总是要去写不同的代码,于是将MaxCompute, Hive, Oracle, MySQL等JDBC连接封装起来,只需要传入不同的参数即可创建一个不同类型的连接池。

连接参数基础类封装

封装了JDBC基础的连接参数,如果不需要这些属性可以继承该类,增加新的属性即可。

@Datapublic class BaseJdbcConnParam implements Serializable {    /**     * driver name     */    private String driverName;    /**     * IP     */    private String ip;    /**     * db server port     */    private Integer port;    /**     * db name     */    private String dbName;    /**     * db connection username     */    private String username;    /**     * db connection password     */    private String password;}

抽象连接工具类封装

功能如下:

/** * @Description 抽象连接工具类父类 * @Author itdl * @Date 2022/08/15 09:54 */public abstract class AbstractConnUtil<P extends BaseJdbcConnParam> {    /**     * connection params     */    protected final P connParam;    /**     * jdbc connection object     */    protected final Connection connection;    /**     * 构造函数, 构造工具类对象     * @param connParam 连接参数     */    public AbstractConnUtil(P connParam) {        this.connParam = connParam;        this.connection = buildConnection();    }    /**     * 构建连接对象     * @return 连接对象     */    protected abstract Connection buildConnection();    /**     * 获取连接     */    public Connection getConnection() {        return connection;    }}

连接池管理

功能如下:

/** * @Description 连接池管理 * @Author itdl * @Date 2022/08/16 09:42 */@Slf4jpublic class DbConnPool<T extends BaseJdbcConnParam> {    /**     * 用于存放连接     */    private final LinkedList<Connection> connPool = new LinkedList<Connection>();    /**     * 最大连接池数量     */    private final Integer maxPoolSize;    private final T connParam;    /**     * 构造函数     * @param connParam 连接参数     * @param maxPoolSize 连接池大小     */    public DbConnPool(T connParam, Integer maxPoolSize)  {        this.maxPoolSize = maxPoolSize;        this.connParam = connParam;        // 初始化连接池        for (int i = 0; i < maxPoolSize; i++) {            connPool.addLast(this.createConnection());        }    }    /**     * 创建数据库连接     * @return 连接     */    private Connection createConnection() {        if (connParam instanceof OracleJdbcConnParam){            final OracleConnUtil util = new OracleConnUtil((OracleJdbcConnParam) connParam);            return util.getConnection();        }        if (connParam instanceof HiveJdbcConnParam){            final HiveConnUtil util = new HiveConnUtil((HiveJdbcConnParam) connParam);            return util.getConnection();        }        if (connParam instanceof MysqlJdbcConnParam){            final MysqlConnUtil util = new MysqlConnUtil((MysqlJdbcConnParam) connParam);            return util.getConnection();        }        if (connParam instanceof MaxComputeJdbcConnParam){            final MaxComputeJdbcUtil util = new MaxComputeJdbcUtil((MaxComputeJdbcConnParam) connParam);            return util.getConnection();        }        throw new BizException(ResultCode.CONN_TYPE_NOT_SUPPORT);    }    /**     * 获取连接     * @return 连接     */    public synchronized Connection getConnection(){        if (connPool.size() == 0){//            throw new BizException(ResultCode.CONN_POOL_EMPTY_ERR);            // 最长等待十分钟            try {                log.info("==========连接池已经空了, 请等待其他线程释放==========");                wait(10 * 60 * 1000);            } catch (InterruptedException e) {                log.info("==========连接池已经空了, 等待了10分钟还没有释放,抛出异常==========");                e.printStackTrace();                throw new BizException(ResultCode.CONN_POOL_EMPTY_ERR);            }        }        // 去除最上面一个连接 如果没有连接了,将会抛出异常        return connPool.removeFirst();    }    /**     * 用完后释放连接     * @param conn 要释放的连接     */    public synchronized void freeConnection(Connection conn){        // 通知连接已经释放        notifyAll();        this.connPool.addLast(conn);    }    /**     * 关闭连接池     */    public synchronized void close(){        for (Connection connection : connPool) {            SqlUtil.close(connection);        }    }}

SQL操作工具类

根据连接对象Connection和数据库房源,封装不同的sql执行。执行SQL核心功能封装。

/** * @Description SQL操作工具类 * @Author itdl * @Date 2022/08/10 17:13 */@Slf4jpublic class SqlUtil {    /**查询mysql表注释sql*/    public static final String SELECT_TABLES_MYSQL = "select table_name, table_comment from information_schema.tables where TABLE_SCHEMA = '%s'";    /**查询MaxCompute表注释sql*/    public static final String SELECT_TABLES_MAX_COMPUTE = "select table_name, table_comment from information_schema.tables where TABLE_SCHEMA = '%s'";    /**查询oracle表注释sql*/    public static final String SELECT_TABLES_ORACLE = "SELECT t2.TABLE_NAME as table_name, t2.COMMENTS as table_comment FROM user_tables t1 inner join user_tab_comments t2 on t1.TABLE_NAME = t2.TABLE_NAME";    /**查询hive表注释sql, 先查询表名,根据表名获取建表语句,正则提取表注释*/    public static final String SELECT_TABLES_HIVE = "show tables";    public static final String SELECT_TABLES_2_HIVE = "describe extended %s";    /**分页数量统计Mysql*/    private static final String SELECT_COUNT_MYSQL = "select count(1) from (%s) z";    /**分页数量统计MaxCompute*/    private static final String SELECT_COUNT_MAX_COMPUTE = "select count(1) from (%s) z;";    /**分页数量统计Hive*/    private static final String SELECT_COUNT_ORACLE = "select count(1) from (%s) z";    /**分页数量统计Oracle*/    private static final String SELECT_COUNT_HIVE = "select count(1) from (%s) z";    /**maxCompute开启全表扫描sql*/    private static final String FULL_SCAN_MAX_COMPUTE = "set odps.sql.allow.fullscan=true;";    /**分页查询sql-Mysql*/    private static final String SELECT_PAGE_MYSQL = "select z.* from (%s) z limit %s, %s";    /**分页查询sql-MaxCompute*/    private static final String SELECT_PAGE_MAX_COMPUTE = "select z.* from (%s) z limit %s, %s;";    /**分页查询sql-Hive*/    private static final String SELECT_PAGE_HIVE = "select * from (select row_number() over () as row_num_01,u.* from (%s) u) mm where mm.row_num_01 between %s and %s";    /**分页查询sql-Oracle*/    private static final String SELECT_PAGE_ORACLE = "select * from (SELECT ROWNUM as row_num_01,z.* from (%s) z) h where h.row_num_01 > %s and h.row_num_01 <= %s";    /**数据库连接*/    private final Connection connection;    /**数据库方言*/    private final Integer dbDialect;    /**支持的方言列表*/    private static final List<Integer> supportDbTypes =            Arrays.asList(DbDialectEnum.ORACLE.getCode(), DbDialectEnum.HIVE.getCode(), DbDialectEnum.MYSQL.getCode(), DbDialectEnum.MAX_COMPUTE.getCode());    public SqlUtil(Connection connection, Integer dbDialect) {        if (!supportDbTypes.contains(dbDialect)){            throw new BizException(ResultCode.CONN_TYPE_NOT_SUPPORT);        }        this.connection = connection;        this.dbDialect = dbDialect;    }    /**     * 根据connection获取所有的表和对应的注释     */    public List<TableMetaInfo> getTables(String schemaName){        List<TableMetaInfo> result = new ArrayList<>();        String sql = "";        switch (this.dbDialect){            case 1:                sql = SELECT_TABLES_ORACLE;                break;            case 2:                sql = SELECT_TABLES_HIVE;                break;            case 3:                if (StringUtils.isBlank(schemaName)){                    throw new BizException(ResultCode.SELECT_TABLES_SCHEMA_NOT_NULL_ERR);                }                sql = String.format(SELECT_TABLES_MYSQL, schemaName);                break;            case 4:                if (StringUtils.isBlank(schemaName)){                    throw new BizException(ResultCode.SELECT_TABLES_SCHEMA_NOT_NULL_ERR);                }                sql = String.format(SELECT_TABLES_MAX_COMPUTE, schemaName);            default:                break;        }        if (StringUtils.isBlank(sql)){            throw new BizException(ResultCode.CONN_TYPE_NOT_SUPPORT);        }        // 执行SQL语句        final List<LinkedHashMap<String, Object>> resultMaps = querySql(sql);        if (ObjectUtils.isEmpty(resultMaps)){            return Lists.newArrayList();        }        // hive单独处理        List<TableMetaInfo> result1 = getHiveTableMetaInfos(result, resultMaps);        if (result1 != null) return result1;        // 转换结果        return resultMaps.stream().map(                m->{                    final TableMetaInfo info = new TableMetaInfo();                    Object tableNameObj = m.get("table_name");                    String tableName = tableNameObj == null ? m.get("TABLE_NAME") == null ? "" : String.valueOf(m.get("TABLE_NAME")) : String.valueOf(tableNameObj);                    Object tableCommentObj = m.get("table_comment");                    String tableComment = tableCommentObj == null ? m.get("TABLE_COMMENT") == null ? "" : String.valueOf(m.get("TABLE_COMMENT")) : String.valueOf(tableCommentObj);                    info.setTableName(tableName);                    info.setComment(tableComment);                    return info;                }        ).collect(Collectors.toList());    }    /**     * 根据schemeName,表名获取字段列表     * @param tableName 一般是数据库 oracle是用户名     */    public List<TableColumnMetaInfo> getColumnsByTableName(String tableName){        try {            List<TableColumnMetaInfo> list = new ArrayList<>();            final DatabaseMetaData metaData = connection.getMetaData();            final ResultSet columns = metaData.getColumns(null, null, tableName, null);            while (columns.next()){                String columnName = columns.getString("COLUMN_NAME");                String remarks = columns.getString("REMARKS");                remarks = StringUtils.isBlank(remarks) ? "" : remarks;                final TableColumnMetaInfo metaInfo = new TableColumnMetaInfo(tableName, columnName, remarks);                list.add(metaInfo);            }            return list;        } catch (SQLException e) {            e.printStackTrace();            return Lists.newArrayList();        }    }    /**     * 执行sql查询     * @param querySql 查询sql     * @return List<Map<String, Object>> 通过LinkedHashMap接受,序列化时可保证顺序一致     */    public List<LinkedHashMap<String, Object>> queryData(String querySql, boolean... fullScan){        Statement statement = null;        ResultSet resultSet = null;        try {            // 创建statement            statement = this.connection.createStatement();            // 执行全表扫描sql            for (boolean b : fullScan) {                if (b){                    statement.execute(FULL_SCAN_MAX_COMPUTE);                    break;                }            }            // 执行查询语句            resultSet = statement.executeQuery(querySql);            // 构建结果返回            return buildListMap(resultSet);        } catch (SQLException e) {            e.printStackTrace();            throw new BizException(ResultCode.SQL_EXEC_ERR);        } finally {            // 关闭resultSet, statement            close(resultSet, statement);        }    }    /**     * 执行sql查询     * @param querySql 查询sql     * @return List<Map<String, Object>>     */    public List<LinkedHashMap<String, Object>> queryData(String querySql, Integer page, Integer size){        Statement statement = null;        ResultSet resultSet = null;        try {            // 1、替换分号            querySql = querySql.replaceAll(";", "");            // 创建statement            statement = this.connection.createStatement();            // 2、格式化SQL            int offset = (page - 1 ) * size;            String execSql = "";            switch (this.dbDialect){                case 1:                    // oracle                    execSql = String.format(SELECT_PAGE_ORACLE, querySql, offset, size);                    break;                case 2:                    // hive                    execSql = String.format(SELECT_PAGE_HIVE, querySql, offset, size);                    break;                case 3:                    // mysql                    execSql = String.format(SELECT_PAGE_MYSQL, querySql, offset, size);                    break;                case 4:                    // maxCompute                    execSql = String.format(SELECT_PAGE_MAX_COMPUTE, querySql, offset, size);                    break;                default:                    break;            }            // maxCompute开启全表扫描            if (DbDialectEnum.MAX_COMPUTE.getCode().equals(this.dbDialect)){                statement.execute(FULL_SCAN_MAX_COMPUTE);            }            log.info("=======>>>执行分页sql为:{}", execSql);            // 执行查询语句            resultSet = statement.executeQuery(execSql);            // 构建结果返回            return buildListMap(resultSet);        } catch (SQLException e) {            e.printStackTrace();            throw new BizException(ResultCode.SQL_EXEC_ERR);        } finally {            // 关闭resultSet, statement            close(resultSet, statement);        }    }    /**     * 执行分页查询     * @param querySql 分页查询sql     * @param page 页码 从1开始 第n页传n     * @param size 每页记录数     * @return 分页查询结果     */    public PageResult<LinkedHashMap<String, Object>> pageQueryMap(String querySql, Integer page, Integer size){        // 1、替换分号        querySql = querySql.replaceAll(";", "");        String countSql = "";        switch (this.dbDialect){            case 1:                // oracle                countSql = String.format(SELECT_COUNT_ORACLE, querySql);                break;            case 2:                // hive                countSql = String.format(SELECT_COUNT_HIVE, querySql);                break;            case 3:                // mysql                countSql = String.format(SELECT_COUNT_MYSQL, querySql);                break;            case 4:                // maxCompute                countSql = String.format(SELECT_COUNT_MAX_COMPUTE, querySql);                break;            default:                break;        }        log.info("=======>>>执行分页统计总数sql为:{}", countSql);        // 查询总数        final List<LinkedHashMap<String, Object>> countMap = queryData(countSql, DbDialectEnum.MAX_COMPUTE.getCode().equals(this.dbDialect));        if (CollectionUtils.isEmpty(countMap)){            return new PageResult<>(0L, new ArrayList<>());        }        long count = 0L;        for (Object value : countMap.get(0).values()) {            count = Long.parseLong(String.valueOf(value));        }        if (count == 0){            return new PageResult<>(0L, new ArrayList<>());        }        // 执行分页查询 开启全表扫描        final List<LinkedHashMap<String, Object>> resultList = queryData(querySql, page, size);        return new PageResult<>(count, resultList);    }    /**     * 执行分页查询     * @param querySql 分页查询sql     * @param page 页码 从1开始 第n页传n     * @param size 每页记录数     * @return 分页查询结果     */    public <T>PageResult<T> pageQuery(String querySql, Integer page, Integer size, Class<T> clazz){        final PageResult<LinkedHashMap<String, Object>> result = pageQueryMap(querySql, page, size);        List<T> rows = new ArrayList<>();        for (LinkedHashMap<String, Object> row : result.getRows()) {            final T t = JSONObject.parseobject(JSONObject.toJSONString(row), clazz);            rows.add(t);        }        return new PageResult<>(result.getTotal(), rows);    }    /**     * 获取hive的表注释     * @param result 结果     * @param resultMaps show tables结果     * @return List<TableMetaInfo>     */    private List<TableMetaInfo> getHiveTableMetaInfos(List<TableMetaInfo> result, List<LinkedHashMap<String, Object>> resultMaps) {        if (dbDialect.equals(DbDialectEnum.HIVE.getCode())){            for (LinkedHashMap<String, Object> resultMap : resultMaps) {                final String tabName = String.valueOf(resultMap.get("tab_name"));                final String descTableCommentSql = String.format(SELECT_TABLES_2_HIVE, tabName);                List<LinkedHashMap<String, Object>> resultMapsComments = querySql(descTableCommentSql);//                col_name -> Detailed Table Information                String comments = resultMapsComments.stream()                        .filter(m -> "Detailed Table Information".equals(m.get("col_name")))                        .map(m -> String.valueOf(m.get("data_type"))).findFirst()                        .orElse("");                comments = ReUtil.get("parameters:\{(?!.*?\().*transient_lastDdlTime.*?comment=(.*?)\}", comments,1);                if (StringUtils.isBlank(comments)) {                    comments = "";                }                if (comments.contains(",")){                    comments = comments.substring(0, comments.lastIndexOf(","));                }                result.add(new TableMetaInfo(tabName, comments));                log.info("===========>>>获取表{}的注释成功:{}", tabName, comments);                resultMapsComments.clear();            }            return result;        }        return null;    }    /**     * 执行SQL查询     * @param sql sql语句     * @return 数据列表,使用LinkedHashMap是为了防止HashMap序列化后导致顺序乱序     */    public List<LinkedHashMap<String, Object>> querySql(String sql){        // 执行sql        Statement statement = null;        ResultSet resultSet = null;        try {            statement = connection.createStatement();            resultSet = statement.executeQuery(sql);            return buildListMap(resultSet);        } catch (SQLException e) {            e.printStackTrace();            throw new BizException(ResultCode.SQL_EXEC_ERR);        }finally {            // 关闭            close(resultSet, statement);        }    }    /**     * 关闭对象 传入多个时注意顺序, 需要先关闭哪个就传在参数前面     * @param objs 对象动态数组     */    public static void close(Object ...objs){        if (objs == null || objs.length == 0){            return;        }        for (Object obj : objs) {            if (obj instanceof Statement){                try {                    ((Statement) obj).close();                }catch (Exception e){                    e.printStackTrace();                }            }            if (obj instanceof ResultSet){                try {                    ((ResultSet) obj).close();                }catch (Exception e){                    e.printStackTrace();                }            }            if (obj instanceof Connection){                try {                    ((Connection) obj).close();                }catch (Exception e){                    e.printStackTrace();                }            }        }    }    /**     * @Description 功能描述:将resultSet构造为List<Map>     * @Author itdl     * @Date 2022/4/18 21:13     * @Param {@link ResultSet} resultSet     * @Return {@link List < Map <String,Object>>}     **/    private List<LinkedHashMap<String, Object>> buildListMap(ResultSet resultSet) throws SQLException {        if (resultSet == null) {            return Lists.newArrayList();        }        List<LinkedHashMap<String, Object>> resultList = new ArrayList<>();        // 获取元数据        ResultSetMetaData metaData = resultSet.getMetaData();        while (resultSet.next()) {            // 获取列数            int columnCount = metaData.getColumnCount();            LinkedHashMap<String, Object> map = new LinkedHashMap<>();            for (int i = 0; i < columnCount; i++) {                String columnName = metaData.getColumnName(i + 1);                // 过滤掉查询的结果包含序号的                if("mm.row_num_01".equalsIgnoreCase(columnName)                        || "row_num_01".equalsIgnoreCase(columnName)){                    continue;                }                // 去除hive查询结果的mm.别名前缀                if (columnName.startsWith("mm.")){                    columnName = columnName.substring(columnName.indexOf(".") + 1);                }                Object object = resultSet.getObject(columnName);                // maxCompute里面的空返回的是使用n                if ("\N".equalsIgnoreCase(String.valueOf(object))) {                    map.put(columnName, "");                } else {                    map.put(columnName, object);                }            }            resultList.add(map);        }        return resultList;    }}

MaxCompute JDBC连接池封装

MaxCompute 已经有了JDBC连接方式 也就是 odbc-jdbc, 最终能够获取一个Connection. 官方文档:
https://help.aliyun.com/document_detail/161246.html

封装MaxCompute JDBC连接参数

/** * @author itdl * @description maxCompute使用JDBC的连接参数 * @date 2022/08/08 10:07 */@Datapublic class MaxComputeJdbcConnParam extends BaseJdbcConnParam{    /**阿里云accessId 相当于用户名 */    private String aliyunAccessId;    /**阿里云accessKey 相当于密码 */    private String aliyunAccessKey;    /** maxcompute_endpoint */    private String endpoint;    /**项目名称*/    private String projectName;}

封装MaxCompute JDBC连接实现类

就是实现父类AbstractConnUtil,实现抽象方法buildConnection

/** * @Description maxCompute JDBC连接实现 * @Author itdl * @Date 2022/08/08 14:26 */@Slf4jpublic class MaxComputeJdbcUtil extends AbstractConnUtil<MaxComputeJdbcConnParam>{    /**JDBC 驱动名称*/    private static final String DRIVER_NAME = "com.aliyun.odps.jdbc.OdpsDriver";    /**     * 构造函数, 构造工具类对象     *     * @param connParam 连接参数     */    public MaxComputeJdbcUtil(MaxComputeJdbcConnParam connParam) {        super(connParam);    }    @Override    protected Connection buildConnection() {        return buildConn();    }    /**     * 创建连接     * @return 数据库连接     */    private Connection buildConn() {        try {            Class.forName(DRIVER_NAME);        } catch (ClassNotFoundException e) {            e.printStackTrace();            throw new BizException(ResultCode.MAX_COMPUTE_DRIVE_LOAD_ERR);        }        try {            Properties dbProperties = new Properties();            dbProperties.put("user", connParam.getAliyunAccessId());            dbProperties.put("password", connParam.getAliyunAccessKey());            dbProperties.put("remarks", "true");            // JDBCURL连接模板            String jdbcUrlTemplate = "jdbc:odps:%s?project=%s&useProjectTimeZone=true";            // 使用驱动管理器连接获取连接            return DriverManager.getConnection(                    String.format(jdbcUrlTemplate, connParam.getEndpoint(), connParam.getProjectName()), dbProperties);        } catch (SQLException e) {            e.printStackTrace();            throw new BizException(ResultCode.CONN_USER_PWD_ERR);        }    }}

连接测试代码一起放在结尾,将会开启多个线程获取连接,然后去获取表名,表注释,字段名,字段注释,传入page, size和普通sql就可以实现分页查询的封装方法

Hive JDBC连接池封装

Hive JDBC连接参数

Hive连接参数封装,除了基础的JDBC所需字段,还需要kerberos相关字段,因为hive开启kerberos认证后,需要使用kertab密钥文件和kbr5.conf配置文件去认证。将会在参数和测试代码中得到重复的体现。

/** * @Description Hive JDBC connection params * @Author itdl * @Date 2022/08/10 16:40 */@Data@EqualsAndHashCode(callSuper = false)public class HiveJdbcConnParam extends BaseJdbcConnParam {    /**     * enable kerberos authentication     */    private boolean enableKerberos;    /**     * principal     */    private String principal;    /**     * kbr5 file path in dick     */    private String kbr5FilePath;    /**     * keytab file path in dick     */    private String keytabFilePath;}

Hive JDBC获取连接实现

Hive获取JDBC连接之后,本来可以从Connection的元数据中获取表的注释,但是获取的中文注释居然是乱码,但是我们Hue上查看表注释又是正常,暂时没找到这种方式如何解决,从而退而求其次,通过表名去获取建表语句,从建表语句中通过正则表达式提取表的注释。

/** * @Description hive connection util * @Author itdl * @Date 2022/08/10 16:52 */@Slf4jpublic class HiveConnUtil extends AbstractConnUtil<HiveJdbcConnParam>{    public HiveConnUtil(HiveJdbcConnParam connParam) {        super(connParam);    }    /**     * 获取连接     * @return 连接     */    public Connection getConnection() {        return connection;    }    @Override    protected Connection buildConnection(){        try {//            Class.forName("org.Apache.hive.jdbc.HiveDriver");            Class.forName(connParam.getDriverName());        } catch (ClassNotFoundException e) {            e.printStackTrace();            throw new BizException(ResultCode.HIVE_DRIVE_LOAD_ERR);        }        // 开启kerberos后需要私钥        // 拼接jdbcUrl        String jdbcUrl = "jdbc:hive2://%s:%s/%s";        String ip = connParam.getIp();        String port = connParam.getPort() + "";        String dbName = connParam.getDbName();        final String username = connParam.getUsername();        final String password = connParam.getPassword();        // is enable kerberos authentication        final boolean enableKerberos = connParam.isEnableKerberos();        // 格式化        Connection connection;        // 获取连接        try {            Properties dbProperties = new Properties();            dbProperties.put("user", username);            dbProperties.put("password", password);            // 加上remark后, 能够获取到标注释 但是会出现中文乱码            dbProperties.put("remarks", "true");            if (!enableKerberos) {                jdbcUrl = String.format(jdbcUrl, ip, port, dbName);                connection = DriverManager.getConnection(jdbcUrl, dbProperties);            } else {                final String principal = connParam.getPrincipal();                final String kbr5FilePath = connParam.getKbr5FilePath();                final String secretFilePath = connParam.getKeytabFilePath();                String format = "jdbc:hive2://%s:%s/%s;principal=%s";                jdbcUrl = String.format(format, ip, port, dbName, principal);                // 使用hadoop安全认证                System.setProperty("JAVA.security.krb5.conf", kbr5FilePath);                System.setProperty("javax.security.auth.useSubjectCredsOnly", "false");                // 解决windows中执行可能出现找不到HADOOP_HOME或hadoop.home.dir问题                // Kerberos认证                org.apache.hadoop.conf.Configuration conf = new org.apache.hadoop.conf.Configuration();                conf.set("hadoop.security.authentication", "Kerberos");                conf.set("keytab.file", secretFilePath);                conf.set("kerberos.principal", principal);                UserGroupInformation.setConfiguration(conf);                try {                    UserGroupInformation.loginUserFromKeytab(username, secretFilePath);                } catch (IOException e) {                    e.printStackTrace();                    throw new BizException(ResultCode.KERBEROS_AUTH_FAIL_ERR);                }                try {                    connection = DriverManager.getConnection(jdbcUrl, dbProperties);                } catch (SQLException e) {                    e.printStackTrace();                    throw new BizException(ResultCode.KERBEROS_AUTH_SUCCESS_GET_CONN_FAIL_ERR);                }            }            log.info("=====>>>获取hive连接成功:username:{},jdbcUrl: {}", username, jdbcUrl);            return connection;        } catch (SQLException e) {            e.printStackTrace();            throw new BizException(ResultCode.HIVE_CONN_USER_PWD_ERR);        } catch (BizException e){            throw e;        }        catch (Exception e) {            e.printStackTrace();            throw new BizException(ResultCode.HIVE_CONN_ERR);        }    }}

Oracle JDBC连接参数封装

只需要继承父类即可

/** * @Description Oracle连接的JDBC参数 * @Author itdl * @Date 2022/08/15 09:50 */public class OracleJdbcConnParam extends BaseJdbcConnParam{    }

Oracle JDBC连接实现类

包括了普通用户的认证和dba用户的认证

/** * @Description Oracle获取jdbc连接工具类 * @Author itdl * @Date 2022/08/15 09:52 */@Slf4jpublic class OracleConnUtil extends AbstractConnUtil<OracleJdbcConnParam> {    /**     * 构造函数, 构造工具类对象     *     * @param connParam 连接参数     */    public OracleConnUtil(OracleJdbcConnParam connParam) {        super(connParam);    }    @Override    protected Connection buildConnection() {        try {            Class.forName("oracle.jdbc.driver.OracleDriver");        } catch (ClassNotFoundException e) {            e.printStackTrace();            throw new BizException(ResultCode.ORACLE_DRIVE_LOAD_ERR);        }        // 拼接jdbcUrl        String jdbcUrl = "jdbc:oracle:thin:@//%s:%s/%s";        final String ip = connParam.getIp();        final String port = connParam.getPort() + "";        final String dbName = connParam.getDbName();        final String username = connParam.getUsername();        final String password = connParam.getPassword();        // 格式化        jdbcUrl = String.format(jdbcUrl, ip, port, dbName);        // 获取连接        Connection connection;        try {            Properties dbProperties = new Properties();            // 用户名 如果是dba,则后面跟了as sysdba            String dba = "as sysdba";            dbProperties.put("password", password);            dbProperties.put("remarks", "true");            if (username.trim().endsWith(dba)) {                dbProperties.put("user", username.trim().substring(0, username.trim().indexOf(dba) - 1));                dbProperties.put("defaultRowPrefetch", "15");                dbProperties.put("internal_logon", "sysdba");                connection = DriverManager.getConnection(jdbcUrl, dbProperties);            } else {                dbProperties.put("user", username);                connection = DriverManager.getConnection(jdbcUrl, dbProperties);            }            log.info("=====>>>获取oracle连接成功:username:{}, jdbcUrl: {}", username, jdbcUrl);            return connection;        } catch (SQLException e) {            e.printStackTrace();            if (e.getMessage().contains("TNS:listener")) {                throw new BizException(ResultCode.CONN_LISTENER_UNKNOWN_ERR);            }            if (e.getMessage().contains("ORA-01017")) {                throw new BizException(ResultCode.CONN_USER_PWD_ERR);            }            if (e.getMessage().contains("IO 错误: Got minus one from a read call")) {                throw new BizException(ResultCode.CONN_CONN_TOO_MANY_ERR);            }            throw new BizException(ResultCode.CONN_UNKNOWN_ERR);        } catch (Exception e) {            throw new BizException(ResultCode.CONN_UNKNOWN_ERR);        }    }}

Mysql JDBC连接池封装

Mysql JDBC连接参数封装

只需要继承父类即可

/** * @Description Mysql连接的JDBC参数 * @Author itdl * @Date 2022/08/15 09:50 */public class MysqlJdbcConnParam extends BaseJdbcConnParam{    }

Mysql JDBC连接实现

需要注意的是连接的属性里面配置useInformationSchema=true,表示可以直接从Connection中获取表和字段的注释。

/** * @Description Mysql获取jdbc连接工具类 * @Author itdl * @Date 2022/08/15 09:52 */@Slf4jpublic class MysqlConnUtil extends AbstractConnUtil<MysqlJdbcConnParam> {    /**     * 构造函数, 构造工具类对象     *     * @param connParam 连接参数     */    public MysqlConnUtil(MysqlJdbcConnParam connParam) {        super(connParam);    }    @Override    protected Connection buildConnection() {        try {            Class.forName("com.mysql.cj.jdbc.Driver");        } catch (ClassNotFoundException e) {            e.printStackTrace();            throw new BizException(ResultCode.MYSQL_DRIVE_LOAD_ERR);        }        // 拼接jdbcUrl        String jdbcUrl = "jdbc:mysql://%s:%s/%s?serverTimezone=Asia/Shanghai&useUnicode=true&characterEncoding=UTF-8";        final String ip = connParam.getIp();        final String port = connParam.getPort() + "";        final String dbName = connParam.getDbName();        final String username = connParam.getUsername();        final String password = connParam.getPassword();        // 格式化        jdbcUrl = String.format(jdbcUrl, ip, port, dbName);        // 获取连接        try {            Properties dbProperties = new Properties();            dbProperties.put("user", username);            dbProperties.put("password", password);            dbProperties.put("remarks", "true");            // 设置可以获取tables remarks信息            dbProperties.setProperty("useInformationSchema", "true");            Connection connection = DriverManager.getConnection(jdbcUrl,dbProperties);            log.info("=====>>>获取mysql连接成功:username:{}, jdbcUrl: {}", username, jdbcUrl);            return connection;        } catch (SQLException e) {            e.printStackTrace();            if (e.getMessage().contains("Unknown database")){                throw new BizException(ResultCode.CONN_UNKNOWN_DB_ERR);            }            throw new BizException(ResultCode.CONN_USER_PWD_ERR);        } catch (Exception e) {            throw new BizException(ResultCode.CONN_UNKNOWN_ERR);        }    }}

测试代码连接各自数据库

@SpringBootTest(classes = DbConnectionDemoApplication.class)@RunWith(value = SpringRunner.class)@Slf4jclass DbConnectionDemoApplicationTests {    private DbConnPool<?> connPool = null;    @Test    public void testMysqlConn() throws InterruptedException {        // 创建连接参数        final MysqlJdbcConnParam connParam = new MysqlJdbcConnParam();        final String ip = "localhost";        final Integer port = 3306;        final String username = "root";        final String password = "root";        final String dbname = "test_db";        // 设置参数        connParam.setDriverName(Driver.class.getName());        connParam.setIp(ip);        connParam.setPort(port);        connParam.setUsername(username);        connParam.setPassword(password);        connParam.setDbName(dbname);        // 创建连接池        connPool = new DbConnPool<>(connParam, 2);        handler01(dbname, DbDialectEnum.MYSQL);        new Thread(() -> handler01(dbname, DbDialectEnum.MYSQL)).start();        new Thread(() -> handler01(dbname, DbDialectEnum.MYSQL)).start();        Thread.sleep(60 * 1000);    }    @Test    public void testOracleConn() throws InterruptedException {        // 创建连接参数        final OracleJdbcConnParam connParam = new OracleJdbcConnParam();        final String ip = "你的Oracle的IP地址";        final Integer port = 1521;        // 如果是admin账号 用户后面+ as sysdba        final String username = "用户名";        final String password = "密码";        final String dbname = "实例/服务名";        // 设置参数        connParam.setDriverName(Driver.class.getName());        connParam.setIp(ip);        connParam.setPort(port);        connParam.setUsername(username);        connParam.setPassword(password);        connParam.setDbName(dbname);        // 创建连接池        connPool = new DbConnPool<>(connParam, 2);        final DbDialectEnum dbDialectEnum = DbDialectEnum.ORACLE;        // 处理操作(oracle的schemaName就是用户名)        handler01(username, dbDialectEnum);        // 新建两个线程获取连接        new Thread(() -> handler01(username, dbDialectEnum)).start();        new Thread(() -> handler01(username, dbDialectEnum)).start();        Thread.sleep(60 * 1000);    }    @Test    public void testHiveConn() throws InterruptedException {        // 创建连接参数        final HiveJdbcConnParam connParam = new HiveJdbcConnParam();        final String ip = "连接的域名";        final Integer port = 10000;        // 如果是admin账号 用户后面+ as sysdba        final String username = "账号@域名";        final String password = "";        final String dbname = "数据库名";        final String principal = "hive/_HOST@域名";        final String kbr5FilePath = "C:\workspace\krb5.conf";        final String keytabFilePath = "C:\workspace\zhouyu.keytab";        // 设置参数        connParam.setDriverName(Driver.class.getName());        connParam.setIp(ip);        connParam.setPort(port);        connParam.setUsername(username);        connParam.setPassword(password);        connParam.setDbName(dbname);        connParam.setEnableKerberos(true);        connParam.setPrincipal(principal);        connParam.setKbr5FilePath(kbr5FilePath);        connParam.setKeytabFilePath(keytabFilePath);        // 创建连接池        connPool = new DbConnPool<>(connParam, 2);        final DbDialectEnum dbDialectEnum = DbDialectEnum.HIVE;        // 处理操作(oracle的schemaName就是用户名)        handler01(username, dbDialectEnum);        // 新建两个线程获取连接        new Thread(() -> handler01(username, dbDialectEnum)).start();        new Thread(() -> handler01(username, dbDialectEnum)).start();        Thread.sleep(10 * 60 * 1000);    }    @Test    public void testMaxComputeConn() throws InterruptedException {        // 创建连接参数        final MaxComputeJdbcConnParam connParam = new MaxComputeJdbcConnParam();        String accessId = "你的阿里云accessId";        String accessKey = "你的阿里云accessKey";        String endpoint = "http://service.cn-chengdu.maxcompute.aliyun.com/api";        String projectName = "项目名=数据库名";        // 设置参数        connParam.setDriverName(Driver.class.getName());        connParam.setAliyunAccessId(accessId);        connParam.setAliyunAccessKey(accessKey);        connParam.setEndpoint(endpoint);        connParam.setProjectName(projectName);        // 创建连接池        connPool = new DbConnPool<>(connParam, 2);        final DbDialectEnum dbDialectEnum = DbDialectEnum.MAX_COMPUTE;        // 处理操作(oracle的schemaName就是用户名)        handler01(projectName, dbDialectEnum);        // 新建两个线程获取连接        new Thread(() -> handler01(projectName, dbDialectEnum)).start();        new Thread(() -> handler01(projectName, dbDialectEnum)).start();        Thread.sleep(60 * 1000);    }    private void handler01(String schemaName, DbDialectEnum dbDialectEnum) {        final Connection connection = connPool.getConnection();        // 构建工具类        final SqlUtil sqlUtil = new SqlUtil(connection, dbDialectEnum.getCode());        // 获取表和注释        final List<TableMetaInfo> tables = sqlUtil.getTables(schemaName);        log.info("===============获取所有表和注释开始===================");        log.info(tables.toString());        log.info("===============获取所有表和注释结束===================");        // 获取字段和注释        final String tableName = tables.get(0).getTableName();        final List<TableColumnMetaInfo> columns = sqlUtil.getColumnsByTableName(tableName);        log.info("===============获取第一个表的字段和注释开始===================");        log.info(columns.toString());        log.info("===============获取第一个表的字段和注释结束===================");        final PageResult<LinkedHashMap<String, Object>> pageResult = sqlUtil.pageQueryMap("select * from " + tableName, 1, 10);        log.info("===============SQL分页查询开始===================");        log.info("总数:{}", pageResult.getTotal());        log.info("记录数:{}", JSONObject.toJSONString(pageResult.getRows()));        log.info("===============SQL分页查询结束===================");        connPool.freeConnection(connection);    }    @After    public void close(){        if (connPool != null){            connPool.close();            log.info("==================连接池成功关闭================");        }    }}

小结

就是为了方便整合第三方数据源做数据源管理时比较重要(若有所需,私信“封装数据源源码”获取源码)。

声明:本站部分内容来自互联网,如有版权侵犯或其他问题请与我们联系,我们将立即删除或处理。
▍相关推荐
更多资讯 >>>