sqoop部署

- 下载

在hadoop家目录下创建一个app目录

[hadoop@hadoop001 ~]$ mkdir app

[hadoop@hadoop001 app]$ pwd
/home/hadoop/app

[hadoop@hadoop001 app]$ wget http://archive.cloudera.com/cdh5/cdh/5/sqoop-1.4.6-cdh5.7.0.tar.gz  --下载sqoop安装包  我们用的是cdh环境,所以下载的也是cdh安装包

- 解压
[hadoop@hadoop001 app]$ tar -zxvf sqoop-1.4.6-cdh5.7.0.tar.gz

[hadoop@hadoop001 app]$ ll
drwxr-xr-x 10 hadoop hadoop 4096 Apr 4 15:10 sqoop-1.4.6-cdh5.7.0
-rw-r--r-- 1 root root 29966286 Apr 3 17:38 sqoop-1.4.6-cdh5.7.0.tar.gz

-配置环境变量

[hadoop@hadoop001 ~]$ vi ~/.bash_profile

export SQOOP_HOME=/home/hadoop/app/sqoop

export PATH=$SQOOP_HOME/bin:$PATH

[hadoop@hadoop001 ~]$ source ~/.bash_profile   

-进入conf目录

[hadoop@hadoop001 ~]$ cd $SQOOP_HOME/conf    

-复制配置文件模板并改名

[hadoop@hadoop001 conf]$ cp sqoop-env-template.sh sqoop-env.sh

-编辑环境信息

[hadoop@hadoop001 conf]$ vi sqoop-env.sh

-解开hadoop和hive的注释,并配置正确的目录

#Set path to where bin/hadoop is available
export HADOOP_COMMON_HOME=/home/hadoop/app/hadoop-2.6.0-cdh5.7.0

#Set path to where hadoop-*-core.jar is available
export HADOOP_MAPRED_HOME=/home/hadoop/app/hadoop-2.6.0-cdh5.7.0

#set the path to where bin/hbase is available
#export HBASE_HOME=

#Set the path to where bin/hive is available
export HIVE_HOME=/home/hadoop/app/hive-1.1.0-cdh5.7.0

导入jdbc驱动jar包

[hadoop@hadoop001 conf]$ cd $SQOOP_HOME/lib

[hadoop@hadoop001 lib]$ wget http://central.maven.org/maven2/mysql/mysql-connector-java/5.1.27/mysql-connector-java-5.1.27.jar

导入json包

[hadoop@hadoop001 lib]$ wget http://www.java2s.com/Code/JarDownload/java-json/java-json.jar.zip

[hadoop@hadoop001 lib]$ unzip java-json.jar.zip

导入hive-exec-**.jar

[hadoop@hadoop001 lib]$ cp $HIVE_HOME/lib/hive-exec-**.jar $SQOOP_HOME/lib

输入sqoop version检查是否安装成功

[hadoop@hadoop001 lib]$ sqoop version
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../hbase does not exist! HBase imports will fail.
Please set $HBASE_HOME to the root of your HBase installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../hcatalog does not exist! HCatalog jobs will fail.
Please set $HCAT_HOME to the root of your HCatalog installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../accumulo does not exist! Accumulo imports will fail.
Please set $ACCUMULO_HOME to the root of your Accumulo installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../zookeeper does not exist! Accumulo imports will fail.
Please set $ZOOKEEPER_HOME to the root of your Zookeeper installation.
19/04/04 20:19:41 INFO sqoop.Sqoop: Running Sqoop version: 1.4.6-cdh5.7.0
Sqoop 1.4.6-cdh5.7.0
git commit id
Compiled by jenkins on Wed Mar 23 11:30:51 PDT 2016

查看mysql库信息
[hadoop@hadoop001 lib]$ sqoop list-databases \

> --connect jdbc:mysql://localhost:3306 \
> --username root --password 123456

Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../hbase does not exist! HBase imports will fail.
Please set $HBASE_HOME to the root of your HBase installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../hcatalog does not exist! HCatalog jobs will fail.
Please set $HCAT_HOME to the root of your HCatalog installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../accumulo does not exist! Accumulo imports will fail.
Please set $ACCUMULO_HOME to the root of your Accumulo installation.
Warning: /home/hadoop/app/sqoop-1.4.6-cdh5.7.0/../zookeeper does not exist! Accumulo imports will fail.
Please set $ZOOKEEPER_HOME to the root of your Zookeeper installation.
19/04/04 20:25:09 INFO sqoop.Sqoop: Running Sqoop version: 1.4.6-cdh5.7.0
19/04/04 20:25:09 WARN tool.BaseSqoopTool: Setting your password on the command-line is insecure. Consider using -P instead.
19/04/04 20:25:09 INFO manager.MySQLManager: Preparing to use a MySQL streaming resultset.
information_schema
mysql
performance_schema
ruoze_d6
ruozedata
test

查看sqoop命令  
[hadoop@hadoop001 lib]$ sqoop help

codegen                   Generate code to interact with database records
create-hive-table       Import a table definition into Hive
eval                           Evaluate a SQL statement and display the results
export                       Export an HDFS directory to a database table
help                          List available commands
import                       Import a table from a database to HDFS
import-all-tables       Import tables from a database to HDFS
import-mainframe     Import datasets from a mainframe server to HDFS
job                           Work with saved jobs
list-databases          List available databases on a server
list-tables                 List available tables in a database
merge                      Merge results of incremental imports
metastore                 Run a standalone Sqoop metastore
version                      Display version information

通过sqoop把数据从mysql导入hive

sqoop import \
--connect jdbc:mysql://localhost:3306/hiveDB \
--username root --password 123456 \
--table DBS \
--hive-import \
--hive-database myhive \
--hive-table hiveDBS \
--fields-terminated-by '\t' \
--lines-terminated-by '\n' \
-m 2

 

 

 

 

 

 

posted @ 2019-04-04 20:40  任重而道远的小蜗牛  阅读(386)  评论(0编辑  收藏  举报