sqoop1.99.4 JAVA API操作
来源:互联网 发布:sql删除有约束字段 编辑:程序博客网 时间:2024/06/13 12:37
sqoop1.99.4 JAVA API操作
如果你是MAVEN项目
1 <dependency>2 <groupId>org.apache.sqoop</groupId>3 <artifactId>sqoop-client</artifactId>4 <version>1.99.4</version>5 </dependency>
如果你是java项目
导入sqoop1.99.4中shell目录下的lib里面全部jar包就行(不用server中的)
HDFS->MYSQL
package org.admln.sqoopOperate;import org.apache.sqoop.client.SqoopClient;import org.apache.sqoop.model.MFromConfig;import org.apache.sqoop.model.MJob;import org.apache.sqoop.model.MLink;import org.apache.sqoop.model.MLinkConfig;import org.apache.sqoop.model.MSubmission;import org.apache.sqoop.model.MToConfig;import org.apache.sqoop.submission.counter.Counter;import org.apache.sqoop.submission.counter.CounterGroup;import org.apache.sqoop.submission.counter.Counters;import org.apache.sqoop.validation.Status;public class HDFSToMysql { public static void main(String[] args) { sqoopTransfer(); } public static void sqoopTransfer() { //初始化 String url = "http://hadoop:12000/sqoop/"; SqoopClient client = new SqoopClient(url); //创建一个源链接 HDFS long fromConnectorId = 1; MLink fromLink = client.createLink(fromConnectorId); fromLink.setName("HDFS connector"); fromLink.setCreationUser("admln"); MLinkConfig fromLinkConfig = fromLink.getConnectorLinkConfig(); fromLinkConfig.getStringInput("linkConfig.uri").setValue("hdfs://hadoop:8020/"); Status fromStatus = client.saveLink(fromLink); if(fromStatus.canProceed()) { System.out.println("创建HDFS Link成功,ID为: " + fromLink.getPersistenceId()); } else { System.out.println("创建HDFS Link失败"); } //创建一个目的地链接 JDBC long toConnectorId = 2; MLink toLink = client.createLink(toConnectorId); toLink.setName("JDBC connector"); toLink.setCreationUser("admln"); MLinkConfig toLinkConfig = toLink.getConnectorLinkConfig(); toLinkConfig.getStringInput("linkConfig.connectionString").setValue("jdbc:mysql://hadoop:3306/hive"); toLinkConfig.getStringInput("linkConfig.jdbcDriver").setValue("com.mysql.jdbc.Driver"); toLinkConfig.getStringInput("linkConfig.username").setValue("hive"); toLinkConfig.getStringInput("linkConfig.password").setValue("hive"); Status toStatus = client.saveLink(toLink); if(toStatus.canProceed()) { System.out.println("创建JDBC Link成功,ID为: " + toLink.getPersistenceId()); } else { System.out.println("创建JDBC Link失败"); } //创建一个任务 long fromLinkId = fromLink.getPersistenceId(); long toLinkId = toLink.getPersistenceId(); MJob job = client.createJob(fromLinkId, toLinkId); job.setName("HDFS to MySQL job"); job.setCreationUser("admln"); //设置源链接任务配置信息 MFromConfig fromJobConfig = job.getFromJobConfig(); fromJobConfig.getStringInput("fromJobConfig.inputDirectory").setValue("/out/aboutyunLog/HiveExport/ipstatistical/data"); //创建目的地链接任务配置信息 MToConfig toJobConfig = job.getToJobConfig(); toJobConfig.getStringInput("toJobConfig.schemaName").setValue("aboutyunlog"); toJobConfig.getStringInput("toJobConfig.tableName").setValue("ipstatistical"); //toJobConfig.getStringInput("fromJobConfig.partitionColumn").setValue("id"); // set the driver config values //MDriverConfig driverConfig = job.getDriverConfig(); //driverConfig.getStringInput("throttlingConfig.numExtractors").setValue("3");//这句还没弄明白 Status status = client.saveJob(job); if(status.canProceed()) { System.out.println("JOB创建成功,ID为: "+ job.getPersistenceId()); } else { System.out.println("JOB创建失败。"); } //启动任务 long jobId = job.getPersistenceId(); MSubmission submission = client.startJob(jobId); System.out.println("JOB提交状态为 : " + submission.getStatus()); while(submission.getStatus().isRunning() && submission.getProgress() != -1) { System.out.println("进度 : " + String.format("%.2f %%", submission.getProgress() * 100)); //三秒报告一次进度 try { Thread.sleep(3000); } catch (InterruptedException e) { e.printStackTrace(); } } System.out.println("JOB执行结束... ..."); System.out.println("Hadoop任务ID为 :" + submission.getExternalId()); Counters counters = submission.getCounters(); if(counters != null) { System.out.println("计数器:"); for(CounterGroup group : counters) { System.out.print("\t"); System.out.println(group.getName()); for(Counter counter : group) { System.out.print("\t\t"); System.out.print(counter.getName()); System.out.print(": "); System.out.println(counter.getValue()); } } } if(submission.getExceptionInfo() != null) { System.out.println("JOB执行异常,异常信息为 : " +submission.getExceptionInfo()); } System.out.println("HDFS通过sqoop传输数据到MySQL统计执行完毕"); }}
MYSQL->HDFS
package org.admln.sqoopOperate;import org.apache.sqoop.client.SqoopClient;import org.apache.sqoop.model.MDriverConfig;import org.apache.sqoop.model.MFromConfig;import org.apache.sqoop.model.MJob;import org.apache.sqoop.model.MLink;import org.apache.sqoop.model.MLinkConfig;import org.apache.sqoop.model.MSubmission;import org.apache.sqoop.model.MToConfig;import org.apache.sqoop.submission.counter.Counter;import org.apache.sqoop.submission.counter.CounterGroup;import org.apache.sqoop.submission.counter.Counters;import org.apache.sqoop.validation.Status;public class MysqlToHDFS { public static void main(String[] args) { sqoopTransfer(); } public static void sqoopTransfer() { //初始化 String url = "http://hadoop:12000/sqoop/"; SqoopClient client = new SqoopClient(url); //创建一个源链接 JDBC long fromConnectorId = 2; MLink fromLink = client.createLink(fromConnectorId); fromLink.setName("JDBC connector"); fromLink.setCreationUser("admln"); MLinkConfig fromLinkConfig = fromLink.getConnectorLinkConfig(); fromLinkConfig.getStringInput("linkConfig.connectionString").setValue("jdbc:mysql://hadoop:3306/hive"); fromLinkConfig.getStringInput("linkConfig.jdbcDriver").setValue("com.mysql.jdbc.Driver"); fromLinkConfig.getStringInput("linkConfig.username").setValue("hive"); fromLinkConfig.getStringInput("linkConfig.password").setValue("hive"); Status fromStatus = client.saveLink(fromLink); if(fromStatus.canProceed()) { System.out.println("创建JDBC Link成功,ID为: " + fromLink.getPersistenceId()); } else { System.out.println("创建JDBC Link失败"); } //创建一个目的地链接HDFS long toConnectorId = 1; MLink toLink = client.createLink(toConnectorId); toLink.setName("HDFS connector"); toLink.setCreationUser("admln"); MLinkConfig toLinkConfig = toLink.getConnectorLinkConfig(); toLinkConfig.getStringInput("linkConfig.uri").setValue("hdfs://hadoop:8020/"); Status toStatus = client.saveLink(toLink); if(toStatus.canProceed()) { System.out.println("创建HDFS Link成功,ID为: " + toLink.getPersistenceId()); } else { System.out.println("创建HDFS Link失败"); } //创建一个任务 long fromLinkId = fromLink.getPersistenceId(); long toLinkId = toLink.getPersistenceId(); MJob job = client.createJob(fromLinkId, toLinkId); job.setName("MySQL to HDFS job"); job.setCreationUser("admln"); //设置源链接任务配置信息 MFromConfig fromJobConfig = job.getFromJobConfig(); fromJobConfig.getStringInput("fromJobConfig.schemaName").setValue("sqoop"); fromJobConfig.getStringInput("fromJobConfig.tableName").setValue("sqoop"); fromJobConfig.getStringInput("fromJobConfig.partitionColumn").setValue("id"); MToConfig toJobConfig = job.getToJobConfig(); toJobConfig.getStringInput("toJobConfig.outputDirectory").setValue("/usr/tmp"); MDriverConfig driverConfig = job.getDriverConfig(); driverConfig.getStringInput("throttlingConfig.numExtractors").setValue("3"); Status status = client.saveJob(job); if(status.canProceed()) { System.out.println("JOB创建成功,ID为: "+ job.getPersistenceId()); } else { System.out.println("JOB创建失败。"); } //启动任务 long jobId = job.getPersistenceId(); MSubmission submission = client.startJob(jobId); System.out.println("JOB提交状态为 : " + submission.getStatus()); while(submission.getStatus().isRunning() && submission.getProgress() != -1) { System.out.println("进度 : " + String.format("%.2f %%", submission.getProgress() * 100)); //三秒报告一次进度 try { Thread.sleep(3000); } catch (InterruptedException e) { e.printStackTrace(); } } System.out.println("JOB执行结束... ..."); System.out.println("Hadoop任务ID为 :" + submission.getExternalId()); Counters counters = submission.getCounters(); if(counters != null) { System.out.println("计数器:"); for(CounterGroup group : counters) { System.out.print("\t"); System.out.println(group.getName()); for(Counter counter : group) { System.out.print("\t\t"); System.out.print(counter.getName()); System.out.print(": "); System.out.println(counter.getValue()); } } } if(submission.getExceptionInfo() != null) { System.out.println("JOB执行异常,异常信息为 : " +submission.getExceptionInfo()); } System.out.println("MySQL通过sqoop传输数据到HDFS统计执行完毕"); }}
别问为什么没有MYSQL和HBASE、HIVE互导的代码
0 0
- sqoop1.99.4 JAVA API操作
- sqoop1.99.4 JAVA API操作
- Sqoop1.99.4 Java API实践
- sqoop1.99.4搭建(hadoop2.4.0)
- sqoop1.99.4安装与应用
- java 操作 zookeeper API
- hbase java api操作
- Java API操作HDFS
- hbase java api操作
- hbase Java Api 操作
- hbase java api操作
- Hbase-Java API操作
- Java API 操作Hbase
- Java API操作hbase
- Java API操作HDFS
- HbaseCURD操作java API
- java api操作 hbase
- Java Api操作Kafka
- android中方便为fragment写入参数的FragmentArgs简介
- 如何避免软件工程中最昂贵错误的发生(代码重写)
- Android自定义DataTimePicker(日期选择器)
- MFC对话框程序设计笔记
- 安全测试---checklist
- sqoop1.99.4 JAVA API操作
- GDB中查看STL容器类的内容
- 我的世界(Minecraft)1.8 mod 制作——为你的mod添加自定义音效
- NSUserDefaults 数据存储路径
- 在EOS项目中禁用BPS服务
- ios工作札记2——UIView层次管理(sendSubviewToBack,bringSubviewToFront)
- LeetCode--Sudoku Solver(DFS)
- DataTable转为json并绑定Easyui的datagrid控件
- Square root