本地excel/csv文件-->hive-->mysql

前提:Hadoop集群、hive配置完成

1、在node01开启hadoop集群 : start-dfs.sh   ,   start-yarn.sh

2、node01 : 开启 metastore 

hive所在路径/hive/bin/hive --service metastore

3、克隆node01-->node01(1),开启 hive

hive所在路径/hive/bin/hive

4、将本地文件上传至虚拟机

 

5、在hive中创建表

create table lingshou_csv (InvoiceNo String,
                         StockCode String,
                         Description String,
                         Quantity String,
                         InvoiceDate String,
                         UnitPrice String,
                         CustomerID String,
                         Country String) 
                         ROW format delimited fields terminated by ',' STORED AS TEXTFILE;

6、本地文件导入hive

load data  local inpath  '/kkb/install/hive/lingshou.csv' into table lingshou_csv;

7、查看是否导入成功

select * from lingshou_csv limit 10;

注意要用 limit 限制 ,否则数据过多等待时间太长

8、在mysql中的node01数据库(hive中对应mysql)中新建表

create table lingshou_csv (InvoiceNo VARCHAR(255),
                         StockCode VARCHAR(255),
                         Description  VARCHAR(255),
                         Quantity  VARCHAR(255),
                         InvoiceDate  VARCHAR(255),
                         UnitPrice  VARCHAR(255),
                         CustomerID  VARCHAR(255),
                         Country  VARCHAR(255)); 
                         ROW format delimited fields terminated by ',' STORED AS TEXTFILE;

9、hive数据表导入mysql

bin/sqoop export \
--connect jdbc:mysql://node01:3306/lingshu_csv \
--username root \
--password 20194023 \
--table testtable \
--num-mappers 1 \
--export-dir /user/hive/warehouse/lingshu_csv \
--input-fields-terminated-by ","

10、完成,可通过可视化工具查看

posted @ 2021-09-30 21:05  第厘  阅读(79)  评论(0编辑  收藏  举报