700字范文,内容丰富有趣,生活中的好帮手!
700字范文 > sqoop2 mysql导入hdfs_Sqoop2入门之导入关系型数据库数据到HDFS上(sqoop2-1.99.4版本)

sqoop2 mysql导入hdfs_Sqoop2入门之导入关系型数据库数据到HDFS上(sqoop2-1.99.4版本)

时间:2019-11-16 09:53:03

相关推荐

sqoop2 mysql导入hdfs_Sqoop2入门之导入关系型数据库数据到HDFS上(sqoop2-1.99.4版本)

sqoop2-1.99.4和sqoop2-1.99.3版本操作略有不同:新版本中使用link代替了老版本的connection,其他使用类似。

sqoop2-1.99.4环境搭建参见:Sqoop2环境搭建

启动sqoop2-1.99.4版本客户端:

$SQOOP2_HOME/bin/sqoop.shclient

set server--host hadoop000 --port 12000 --webapp sqoop

查看所有connector:

show connector --all

2connector(s) to show:

Connector withid 1:

Name: hdfs-connector

Class: org.apache.sqoop.connector.hdfs.HdfsConnector

Version:1.99.4-cdh5.3.0Connector withid 2:

Name: generic-jdbc-connector

Class: org.apache.sqoop.connector.jdbc.GenericJdbcConnector

Version:1.99.4-cdh5.3.0

查询所有link:

show link

删除指定link:

delete link --lid x

查询所有job:

show job

删除指定job:

delete job --jid 1

创建generic-jdbc-connector类型的connector

create link --cid 2Name: First Link

JDBC Driver Class: com.mysql.jdbc.Driver

JDBC Connection String: jdbc:mysql://hadoop000:3306/hive

Username: root

Password:****JDBC Connection Properties:

There are currently0 values inthe map:

entry# protocol=tcp

There are currently1 values inthe map:

protocol=tcp

entry#

New link was successfully created with validation status OK and persistentid 3

show link

+----+-------------+-----------+---------+

| Id | Name | Connector | Enabled |

+----+-------------+-----------+---------+

| 3 | First Link | 2 | true |

+----+-------------+-----------+---------+

创建hdfs-connector类型的connector:

create link -cid 1Name: Second Link

HDFS URI: hdfs://hadoop000:8020

New link was successfully created with validation status OK and persistent id 4

show link+----+-------------+-----------+---------+

| Id | Name | Connector | Enabled |

+----+-------------+-----------+---------+

| 3 | First Link | 2 | true |

| 4 | Second Link | 1 | true |

+----+-------------+-----------+---------+

show link -all2link(s) to show:

link withid 3 and name First Link (Enabled: true, Created by null at 15-2-2 ??11:28, Updated by null at 15-2-2 ??11:28)

Using Connectorid 2Link configuration

JDBC Driver Class: com.mysql.jdbc.Driver

JDBC Connection String: jdbc:mysql://hadoop000:3306/hive

Username: root

Password:

JDBC Connection Properties:

protocol=tcp

link withid 4 and name Second Link (Enabled: true, Created by null at 15-2-2 ??11:32, Updated by null at 15-2-2 ??11:32)

Using Connectorid 1Link configuration

HDFS URI: hdfs://hadoop000:8020

根据connector id创建job:

create job -f 3 -t 4Creating jobfor links with from id 3 and to id 4Please fill following values to create new jobobjectName: Sqoopy

From database configuration

Schema name: hive

Table name: TBLS

Table SQL statement:

Table column names:

Partition column name:

Null value allowedforthe partition column:

Boundary query:

ToJob configuration

Output format:0: TEXT_FILE1: SEQUENCE_FILE

Choose:0Compression format:0: NONE1: DEFAULT2: DEFLATE3: GZIP4: BZIP25: LZO6: LZ47: SNAPPY8: CUSTOM

Choose:0Custom compression format:

Output directory: hdfs://hadoop000:8020/sqoop2/tbls_import_demo_sqoop1.99.4

Throttling resources

Extractors:

Loaders:

New job was successfully created with validation status OK and persistentid 2

查询所有job:

show job+----+--------+----------------+--------------+---------+

| Id | Name | From Connector | To Connector | Enabled |

+----+--------+----------------+--------------+---------+

| 2 | Sqoopy | 2 | 1 | true |

+----+--------+----------------+--------------+---------+

启动指定的job: 该job执行完后查看HDFS上的文件(hdfs fs -lshdfs://hadoop000:8020/sqoop2/tbls_import_demo_sqoop1.99.4/)

start job --jid 2

查看指定job的执行状态:

status job --jid 2

停止指定的job:

stop job --jid 2

在start job(如:start job --jid 2)时常见错误:

Exception has occurred during processing command

Exception: org.mon.SqoopException Message: CLIENT_0001:Server has returned exception

在sqoop客户端设置查看job详情:

set option --name verbose --value trueshow job--jid 2

本内容不代表本网观点和政治立场,如有侵犯你的权益请联系我们处理。
网友评论
网友评论仅供其表达个人看法,并不表明网站立场。