Hive数据导出导入操作

1:导入数据进入到Hive表:
LOAD DATA [LOCAL] INPATH 'filepath' [OVERWRITE] INTO TABLE tablename [PARTITION (partcol1=val1, partcol2=val2 ...)]

实例:
1)加载本地数据到Hive表:
load data local inpath '/home/wql/app/hData/emp.txt' into table emp_partition partition(month='20190307');

2)加载HDFS数据到Hive表:
load data inpath '/user/hive/hData/emp_ext2/emp.txt' overwrite into table emp;

3)创建表是通过insert加载
create table default.emp_cli like emp;
insert into table emp_cli select * from emp;

4)创建表的时候通过LOCATION指定加载

2:导出数据
1)直接导出数据:
insert overwrite local directory '/home/wql/app/hData/hive_exp_emp' select * from emp;
2)定义导出数据格式:
insert overwrite local directory '/home/wql/app/hData/hive_exp_emp2' row format delimited fields terminated by '\t' collection items terminated by '\n' select * from emp;
3)hive -e "select * from emp;" > /home/wql/app/hData/hive_exp_emp.txt
4)将数据保存在HDFS文件系统上:
insert overwrite directory '/user/beifeng/hive/warehouse/hive_exp_emp' row format delimited fields terminated by '\t' collection items terminated by '\n' select * from emp;

3:export
将hive表的数据导入到外部
EXPORT TABLE tablename [PARTITION (part_column="value"[, ...])]
TO 'export_target_path' [ FOR replication('eventid') ]
实例:
export table emp to '/user/beifeng/hive/warehouse/export/exp_emp';

4:import
IMPORT [[EXTERNAL] TABLE new_or_original_tablename [PARTITION (part_column="value"[, ...])]]
FROM 'source_path'
[LOCATION 'import_target_path']

create table database_name.emp like default.emp;
import table spark.emp from '/user/beifeng/hive/warehouse/export/exp_emp';

猜你喜欢

转载自blog.csdn.net/wql_zero/article/details/88396019