sqoop2-1.99.4和sqoop2-1.99.3版本操作略有不同:新版本中使用link代替了老版本的connection,其他使用类似。
sqoop2-1.99.4环境搭建参见:Sqoop2环境搭建
启动sqoop2-1.99.4版本客户端:
$SQOOP2_HOME/bin/sqoop.shclient
set server--host hadoop000 --port 12000 --webapp sqoop
查看所有connector:
show connector --all
2connector(s) to show:
Connector withid 1:
Name: hdfs-connector
Class: org.apache.sqoop.connector.hdfs.HdfsConnector
Version:1.99.4-cdh5.3.0Connector withid 2:
Name: generic-jdbc-connector
Class: org.apache.sqoop.connector.jdbc.GenericJdbcConnector
Version:1.99.4-cdh5.3.0
查询所有link:
show link
删除指定link:
delete link --lid x
查询所有job:
show job
删除指定job:
delete job --jid 1
创建generic-jdbc-connector类型的connector
create link --cid 2Name: First Link
JDBC Driver Class: com.mysql.jdbc.Driver
JDBC Connection String: jdbc:mysql://hadoop000:3306/hive
Username: root
Password:****JDBC Connection Properties:
There are currently0 values inthe map:
entry# protocol=tcp
There are currently1 values inthe map:
protocol=tcp
entry#
New link was successfully created with validation status OK and persistentid 3
show link
+----+-------------+-----------+---------+
| Id | Name | Connector | Enabled |
+----+-------------+-----------+---------+
| 3 | First Link | 2 | true |
+----+-------------+-----------+---------+
创建hdfs-connector类型的connector:
create link -cid 1Name: Second Link
HDFS URI: hdfs://hadoop000:8020
New link was successfully created with validation status OK and persistent id 4
show link+----+-------------+-----------+---------+
| Id | Name | Connector | Enabled |
+----+-------------+-----------+---------+
| 3 | First Link | 2 | true |
| 4 | Second Link | 1 | true |
+----+-------------+-----------+---------+
show link -all2link(s) to show:
link withid 3 and name First Link (Enabled: true, Created by null at 15-2-2 ??11:28, Updated by null at 15-2-2 ??11:28)
Using Connectorid 2Link configuration
JDBC Driver Class: com.mysql.jdbc.Driver
JDBC Connection String: jdbc:mysql://hadoop000:3306/hive
Username: root
Password:
JDBC Connection Properties:
protocol=tcp
link withid 4 and name Second Link (Enabled: true, Created by null at 15-2-2 ??11:32, Updated by null at 15-2-2 ??11:32)
Using Connectorid 1Link configuration
HDFS URI: hdfs://hadoop000:8020
根据connector id创建job:
create job -f 3 -t 4Creating jobfor links with from id 3 and to id 4Please fill following values to create new jobobjectName: Sqoopy
From database configuration
Schema name: hive
Table name: TBLS
Table SQL statement:
Table column names:
Partition column name:
Null value allowedforthe partition column:
Boundary query:
ToJob configuration
Output format:0: TEXT_FILE1: SEQUENCE_FILE
Choose:0Compression format:0: NONE1: DEFAULT2: DEFLATE3: GZIP4: BZIP25: LZO6: LZ47: SNAPPY8: CUSTOM
Choose:0Custom compression format:
Output directory: hdfs://hadoop000:8020/sqoop2/tbls_import_demo_sqoop1.99.4
Throttling resources
Extractors:
Loaders:
New job was successfully created with validation status OK and persistentid 2
查询所有job:
show job+----+--------+----------------+--------------+---------+
| Id | Name | From Connector | To Connector | Enabled |
+----+--------+----------------+--------------+---------+
| 2 | Sqoopy | 2 | 1 | true |
+----+--------+----------------+--------------+---------+
启动指定的job: 该job执行完后查看HDFS上的文件(hdfs fs -lshdfs://hadoop000:8020/sqoop2/tbls_import_demo_sqoop1.99.4/)
start job --jid 2
查看指定job的执行状态:
status job --jid 2
停止指定的job:
stop job --jid 2
在start job(如:start job --jid 2)时常见错误:
Exception has occurred during processing command
Exception: org.mon.SqoopException Message: CLIENT_0001:Server has returned exception
在sqoop客户端设置查看job详情:
set option --name verbose --value trueshow job--jid 2