Hive 自定义UDF函数实现日期格式化和字段AES加密

Hive 自定义UDF函数实现日期格式化和字段AES加密

  • 自定义日期格式化UDF函数
  • 自定义字段AES加密函数
  • 函数的临时注册和永久注册
  • 测试UDF函数使用
  • 项目pom.xml

自定义日期格式化UDF函数

  • 首先自定义日期解析转换的工具类
import org.apache.commons.lang3.StringUtils;import java.text.ParseException;import java.text.SimpleDateFormat;import java.util.Calendar;import java.util.Date;/** * 日期解析转换工具类 * @author root */public class DateUtils {    private static String[] PARSE_PATTERNS = {"yyyy-MM-dd", "yyyy年MM月dd日",            "yyyy-MM-dd HH:mm:ss", "yyyy-MM-dd HH:mm", "yyyy/MM/dd",            "yyyy/MM/dd HH:mm:ss", "yyyy/MM/dd HH:mm", "yyyyMMdd","yyyyMMdd HH:mm:ss"};    public static String parseDate(String string, String pattern) {        if (string == null) {            return null;        }        try {            org.apache.commons.lang3.time.DateUtils.parseDate(string, pattern);            return pattern;        } catch (ParseException e) {            return null;        }    }    /**     * 输入日期统一转换为yyyy-MM-dd 格式     * parseDate 自动判断日期类型     * @param rawDate     * @return     */    public static String unionformDate(String rawDate) {        String result = "";        for (String parsePattern : PARSE_PATTERNS) {            try {                String s = parseDate(rawDate, parsePattern);                if (StringUtils.isNotBlank(s)) {                    SimpleDateFormat format = new SimpleDateFormat(s);                    Date parse = format.parse(rawDate);                    SimpleDateFormat format2 = new SimpleDateFormat("yyyy-MM-dd");                    result = format2.format(parse);                }            } catch (Exception e) {                e.printStackTrace();            }        }        return result;    }    /**     * 输入日期统一转换为目标日期格式 格式     * parseDate 自动判断日期类型     * @param rawDate     * @return     */    public static String unionformDateWithPattern(String rawDate, String resultPattern) {        String result = "";        for (String parsePattern : PARSE_PATTERNS) {            try {                String s = parseDate(rawDate, parsePattern);                if (StringUtils.isNotBlank(s)) {                    SimpleDateFormat format = new SimpleDateFormat(s);                    Date parse = format.parse(rawDate);                    SimpleDateFormat format2 = new SimpleDateFormat(resultPattern);                    result = format2.format(parse);                }            } catch (Exception e) {                e.printStackTrace();            }        }        return result;    }}
  • 自定义日期转换UDF函数
import com.miminglamp.utils.DateUtils;import org.apache.hadoop.hive.ql.exec.UDF;/** * 日期统一UDF函数 */public class DateUnionUDF extends UDF {    /**     * 传递原始日期和format     *     * @param singleDate     * @param format     * @return     */    public String evaluate(String singleDate, String format) {        if ("null".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if ("NULL".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if ("".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if (singleDate != null) {            String result = DateUtils.unionformDateWithPattern(singleDate, format);            return result;        } else {            return "9999-99-99";        }    }    /**     * 不传递format默认为 yyyy-MM-dd格式     *     * @param singleDate     * @return     */    public String evaluate(String singleDate) {        if ("null".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if ("NULL".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if ("".equals(singleDate)) {            singleDate = "9999-99-99 00:00:00";        }        if (singleDate != null) {            String result = DateUtils.unionformDateWithPattern(singleDate, "yyyy-MM-dd");            return result;        } else {            return "9999-99-99";        }    }    public static void main(String[] args) {        DateUnionUDF dateUnionUDF = new DateUnionUDF();        System.out.println(dateUnionUDF.evaluate("2020/12/12"));        System.out.println(dateUnionUDF.evaluate("2020/12/12", "yyyy-MM-dd HH:mm:ss"));    }}

自定义字段AES加密函数

  • 首先创建AES加密的工具类
import sun.misc.BASE64Decoder;import sun.misc.BASE64Encoder;import javax.crypto.*;import javax.crypto.spec.SecretKeySpec;import java.io.IOException;import java.nio.charset.StandardCharsets;import java.security.InvalidKeyException;import java.security.NoSuchAlgorithmException;import java.security.SecureRandom;import java.util.Scanner;/* * AES对称加密和解密 */public class AES {    /*     * 加密     * 1.构造密钥生成器     * 2.根据ecnodeRules规则初始化密钥生成器     * 3.产生密钥     * 4.创建和初始化密码器     * 5.内容加密     * 6.返回字符串     */    public static String AESEncode(String encodeRules,String content){        try {            //1.构造密钥生成器,指定为AES算法,不区分大小写            KeyGenerator keygen=KeyGenerator.getInstance("AES");            //2.根据ecnodeRules规则初始化密钥生成器            //生成一个128位的随机源,根据传入的字节数组            SecureRandom random = SecureRandom.getInstance("SHA1PRNG");            random.setSeed(encodeRules.getBytes());            keygen.init(128, random);            //3.产生原始对称密钥            SecretKey original_key=keygen.generateKey();            //4.获得原始对称密钥的字节数组            byte [] raw=original_key.getEncoded();            //5.根据字节数组生成AES密钥            SecretKey key=new SecretKeySpec(raw, "AES");            //6.根据指定算法AES自成密码器            Cipher cipher=Cipher.getInstance("AES");            //7.初始化密码器,第一个参数为加密(Encrypt_mode)或者解密解密(Decrypt_mode)操作,第二个参数为使用的KEY            cipher.init(Cipher.ENCRYPT_MODE, key);            //8.获取加密内容的字节数组(这里要设置为utf-8)不然内容中如果有中文和英文混合中文就会解密为乱码            byte [] byte_encode=content.getBytes(StandardCharsets.UTF_8);            //9.根据密码器的初始化方式--加密:将数据加密            byte [] byte_AES=cipher.doFinal(byte_encode);            //10.将加密后的数据转换为字符串            //这里用Base64Encoder中会找不到包            //解决办法:            //在项目的Build path中先移除JRE System Library,再添加库JRE System Library,重新编译后就一切正常了。            //11.将字符串返回            return new BASE64Encoder().encode(byte_AES);        } catch (NoSuchAlgorithmException | NoSuchPaddingException | InvalidKeyException | IllegalBlockSizeException | BadPaddingException e) {            e.printStackTrace();        }        //如果有错就返加nulll        return null;    }    /*     * 解密     * 解密过程:     * 1.同加密1-4步     * 2.将加密后的字符串反纺成byte[]数组     * 3.将加密内容解密     */    public static String AESDncode(String encodeRules,String content){        try {            //1.构造密钥生成器,指定为AES算法,不区分大小写            KeyGenerator keygen=KeyGenerator.getInstance("AES");            //2.根据ecnodeRules规则初始化密钥生成器            //生成一个128位的随机源,根据传入的字节数组            SecureRandom random = SecureRandom.getInstance("SHA1PRNG");            random.setSeed(encodeRules.getBytes());            keygen.init(128, random);            //3.产生原始对称密钥            SecretKey original_key=keygen.generateKey();            //4.获得原始对称密钥的字节数组            byte [] raw=original_key.getEncoded();            //5.根据字节数组生成AES密钥            SecretKey key=new SecretKeySpec(raw, "AES");            //6.根据指定算法AES自成密码器            Cipher cipher=Cipher.getInstance("AES");            //7.初始化密码器,第一个参数为加密(Encrypt_mode)或者解密(Decrypt_mode)操作,第二个参数为使用的KEY            cipher.init(Cipher.DECRYPT_MODE, key);            //8.将加密并编码后的内容解码成字节数组            byte [] byte_content= new BASE64Decoder().decodeBuffer(content);            /*             * 解密             */            byte [] byte_decode=cipher.doFinal(byte_content);            return new String(byte_decode, StandardCharsets.UTF_8);        } catch (NoSuchAlgorithmException | NoSuchPaddingException | InvalidKeyException | IOException | IllegalBlockSizeException | BadPaddingException e) {            e.printStackTrace();        }        //如果有错就返加nulll        return null;    }    public static void main(String[] args) {        Scanner scanner=new Scanner(System.in);        /*         * 加密         */        String encodeRules = "cdp2021";        System.out.println("使用AES对称加密,请输入加密的规则");        System.out.println("请输入要加密的内容:");        String content = scanner.next();        System.out.println("根据输入的规则" encodeRules "加密后的密文是:"  AESEncode(encodeRules, content));        /*         * 解密         */        System.out.println("使用AES对称解密,请输入加密的规则:(须与加密相同)");        System.out.println("请输入要解密的内容(密文):");        content = scanner.next();        System.out.println("根据输入的规则" encodeRules "解密后的明文是:"  AESDncode(encodeRules, content));    }}
  • 创建AES UDF函数
import com.miminglamp.aes.AES;import org.apache.hadoop.hive.ql.exec.UDF;public class AESUDF extends UDF {    String password = "CDP2021";    public String evaluate(String type, String content) throws Exception {        if (content == null) return null;        if (!type.equals("encode") && !type.equals("decode")) {            throw new Exception("Parmeter one is needed encode/decode");        }        if (type.equals("encode")) {            //进行加密            return AES.AESEncode(password, content);        } else {            //进行解密            return AES.AESDncode(password, content);        }    }    public static void main(String[] args) {        try {            System.out.println(new AESUDF().evaluate("encode", "test"));            System.out.println(new AESUDF().evaluate("decode", "2YDc41 Y7OIKBgfqNMwDD9koGoxAy52jGiRdJC8A/ o="));        } catch (Exception e) {            e.printStackTrace();        }    }}

函数的临时注册和永久注册

- 永久注册:上传Jar包到HDFS的指定路径CREATE FUNCTION 数据库名.函数名 AS '包名.类名' USING JAR 'hdfs:///path/xxxx.jar';- 删除永久函数:drop function dev.dateunion;临时注册:add jar /path/xx.jar(存储在本地磁盘)- 临时注册UDF函数(hive会话生效)create temporary function 函数名 as '包名.类名';- 删除临时函数:drop temporary function 数据库名.函数名;

测试UDF函数使用

# 永久UDF函数CREATE FUNCTION dev.dateunion AS 'com.miminglamp.udf.DateUnionUDF' USING JAR 'hdfs:///opt/kezhen/hive/udf/HiveUDF-1.0.jar';# 永久UDF函数CREATE FUNCTION dev.aesencodes AS 'com.miminglamp.udf.AESUDF' USING JAR 'hdfs:///opt/kezhen/hive/udf/HiveUDF-1.0.jar';# 使用默认格式yyyy-MM-dd 格式select id,dev.dateunion(times) times from app_tmp.hive_udf;# 传入目标格式select id,dev.dateunion(times,'yyyy-MM-dd HH:mm:ss') times from app_tmp.hive_udf;# 使用AES加密select id,dev.aesencodes('encode',times) times from app_tmp.hive_udf;# 使用AES解密select id, dev.aesencodes('decode',times) times from app_tmp.aesencode;

项目pom.xml

创建普通的IDEA的maven-quickstart工程即可,创建工程参考:https://blog.csdn.net/qq_43081842/article/details/105294308

最后附上Pom.xml

    <properties>        <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>        <maven.compiler.source>1.7</maven.compiler.source>        <maven.compiler.target>1.7</maven.compiler.target>        <jdk.version>1.8</jdk.version>        <scala.version>2.11.8</scala.version>        <scala.compat.version>2.11</scala.compat.version>        <hadoop.version>3.0.0-cdh6.0.1</hadoop.version>        <spark.version>2.2.0-cdh6.0.1</spark.version>        <hive.version>2.1.1-cdh6.0.1</hive.version>    </properties>    <repositories>        <repository>            <id>cloudera</id>            <url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>        </repository>        <repository>            <id>aliyun</id>            <url>http://maven.aliyun.com/nexus/content/groups/public/</url>        </repository>        <repository>            <id>jboss</id>            <url>http://repository.jboss.com/nexus/content/groups/public</url>        </repository>    </repositories>    <dependencies>        <dependency>            <groupId>junit</groupId>            <artifactId>junit</artifactId>            <version>4.11</version>            <scope>test</scope>        </dependency>        <dependency>            <groupId>redis.clients</groupId>            <artifactId>jedis</artifactId>            <version>3.0.1</version>            <!--            <scope>provided</scope>-->        </dependency>        <dependency>            <groupId>org.apache.spark</groupId>            <artifactId>spark-core_2.11</artifactId>            <version>${spark.version}</version>            <!--            <scope>provided</scope>-->        </dependency>        <dependency>            <groupId>org.apache.spark</groupId>            <artifactId>spark-sql_2.11</artifactId>            <version>${spark.version}</version>            <!--            <scope>provided</scope>-->        </dependency>        <dependency>            <groupId>com.alibaba</groupId>            <artifactId>fastjson</artifactId>            <version>1.2.8.sec06</version>        </dependency>        <dependency>            <groupId>org.apache.spark</groupId>            <artifactId>spark-hive_2.11</artifactId>            <version>${spark.version}</version>            <scope>provided</scope>        </dependency>        <dependency>            <groupId>org.apache.commons</groupId>            <artifactId>commons-lang3</artifactId>            <version>3.3.2</version>        </dependency>        <dependency>            <groupId>org.apache.httpcomponents</groupId>            <artifactId>httpcore</artifactId>            <version>4.4.3</version>        </dependency>        <dependency>            <groupId>org.apache.httpcomponents</groupId>            <artifactId>httpclient</artifactId>            <version>4.5.1</version>        </dependency>        <dependency>            <groupId>org.slf4j</groupId>            <artifactId>slf4j-api</artifactId>            <version>1.7.12</version>        </dependency>        <dependency>            <groupId>log4j</groupId>            <artifactId>log4j</artifactId>            <version>1.2.17</version>        </dependency>        <dependency>            <groupId>org.slf4j</groupId>            <artifactId>slf4j-log4j12</artifactId>            <version>1.7.12</version>        </dependency>        <!--        <dependency>-->        <!--            <groupId>org.apache.spark</groupId>-->        <!--            <artifactId>spark-hive-thriftserver_2.11</artifactId>-->        <!--            <version>2.2.0</version>-->        <!--        </dependency>-->        <!--mysql依赖的jar包-->        <dependency>            <groupId>mysql</groupId>            <artifactId>mysql-connector-java</artifactId>            <version>5.1.35</version>        </dependency>        <dependency>            <groupId>org.apache.hadoop</groupId>            <artifactId>hadoop-client</artifactId>            <version>${hadoop.version}</version>        </dependency>        <dependency>            <groupId>org.apache.hadoop</groupId>            <artifactId>hadoop-common</artifactId>            <version>${hadoop.version}</version>        </dependency>        <dependency>            <groupId>org.apache.hadoop</groupId>            <artifactId>hadoop-hdfs</artifactId>            <version>${hadoop.version}</version>        </dependency>        <dependency>            <groupId>org.apache.hadoop</groupId>            <artifactId>hadoop-mapreduce-client-core</artifactId>            <version>${hadoop.version}</version>        </dependency>        <!--hive 依赖-->        <dependency>            <groupId>org.apache.hive</groupId>            <artifactId>hive-exec</artifactId>            <version>${hive.version}</version>        </dependency>    </dependencies>

来源:https://www.icode9.com/content-4-813251.html

(0)

相关推荐