用MapReduce实现矩阵乘法
来源:互联网 发布:大数据都要学什么 编辑:程序博客网 时间:2024/04/29 23:37
主要介绍Hadoop家族产品,常用的项目包括Hadoop, Hive, Pig, HBase, Sqoop, Mahout, Zookeeper, Avro, Ambari, Chukwa,新增加的项目包括,YARN, Hcatalog, Oozie, Cassandra, Hama, Whirr, Flume, Bigtop, Crunch, Hue等。
从2011年开始,中国进入大数据风起云涌的时代,以Hadoop为代表的家族软件,占据了大数据处理的广阔地盘。开源界及厂商,所有数据软件,无一不向Hadoop靠拢。Hadoop也从小众的高富帅领域,变成了大数据开发的标准。在Hadoop原有技术基础之上,出现了Hadoop家族产品,通过“大数据”概念不断创新,推出科技进步。
作为IT界的开发人员,我们也要跟上节奏,抓住机遇,跟着Hadoop一起雄起!
关于作者:
- 张丹(Conan), 程序员Java,R,PHP,Javascript
- weibo:@Conan_Z
- blog: http://blog.fens.me
- email: bsspirit@gmail.com
转载请注明出处:
http://blog.fens.me/hadoop-mapreduce-matrix/
前言
MapReduce打开了并行计算的大门,让我们个人开发者有了处理大数据的能力。但想用好MapReduce,把原来单机算法并行化,也不是一件容易事情。很多的时候,我们需要从单机算法能否矩阵化去思考,所以矩阵操作就变成了算法并行化的基础。
像推荐系统的协同过滤算法,就是基于矩阵思想实现MapReduce并行化。
目录
- 矩阵介绍
- 矩阵乘法的R语言计算
- 矩阵乘法的MapReduce计算
- 稀疏矩阵乘法的MapReduce计算
1. 矩阵介绍
矩阵: 数学上,一个m×n的矩阵是一个由m行n列元素排列成的矩形阵列。矩阵里的元素可以是数字、符号或数学式。以下是一个由6个数字符素构成的2行3列的矩阵:
1 2 34 5 6
矩阵加法
大小相同(行数列数都相同)的矩阵之间可以相互加减,具体是对每个位置上的元素做加减法。
举例:两个矩阵的加法
1 3 1 + 0 0 5 = 1+0 3+0 1+5 = 1 3 61 0 0 7 5 0 1+7 0+5 0+0 8 5 0
矩阵乘法
两个矩阵可以相乘,当且仅当第一个矩阵的列数等于第二个矩阵的行数。矩阵的乘法满足结合律和分配律,但不满足交换律。
举例:两个矩阵的乘法
1 0 2 * 3 1 = (1*3+0*2+2*1) (1*1+0*1+2*0) = 5 1-1 3 1 2 1 (-1*3+3*2+1*1) (-1*1+3*1+1*0) 4 2 1 0
2. 矩阵乘法的R语言计算
> m1<-matrix(c(1,0,2,-1,3,1),nrow=2,byrow=TRUE);m1 [,1] [,2] [,3][1,] 1 0 2[2,] -1 3 1> m2<-matrix(c(3,1,2,1,1,0),nrow=3,byrow=TRUE);m2 [,1] [,2][1,] 3 1[2,] 2 1[3,] 1 0> m3<-m1 %*% m2;m3 [,1] [,2][1,] 5 1[2,] 4 2
由R语言实现矩阵的乘法是非常简单的。
3. 矩阵乘法的MapReduce计算
算法实现思路:
- 新建2个矩阵数据文件:m1.csv, m2.csv
- 新建启动程序:MainRun.java
- 新建MR程序:MartrixMultiply.java
1).新建2个矩阵数据文件m1.csv, m2.csv
m1.csv
1,0,2-1,3,1
m2.csv
3,12,11,0
3).新建启动程序:MainRun.java
启动程序
package org.conan.myhadoop.matrix;import java.util.HashMap;import java.util.Map;import java.util.regex.Pattern;import org.apache.hadoop.mapred.JobConf;public class MainRun { public static final String HDFS = "hdfs://192.168.1.210:9000"; public static final Pattern DELIMITER = Pattern.compile("[\t,]"); public static void main(String[] args) { martrixMultiply(); } public static void martrixMultiply() { Map<String, String> path = new HashMap<String, String>(); path.put("m1", "logfile/matrix/m1.csv");// 本地的数据文件 path.put("m2", "logfile/matrix/m2.csv"); path.put("input", HDFS + "/user/hdfs/matrix");// HDFS的目录 path.put("input1", HDFS + "/user/hdfs/matrix/m1"); path.put("input2", HDFS + "/user/hdfs/matrix/m2"); path.put("output", HDFS + "/user/hdfs/matrix/output"); try { MartrixMultiply.run(path);// 启动程序 } catch (Exception e) { e.printStackTrace(); } System.exit(0); } public static JobConf config() {// Hadoop集群的远程配置信息 JobConf conf = new JobConf(MainRun.class); conf.setJobName("MartrixMultiply"); conf.addResource("classpath:/hadoop/core-site.xml"); conf.addResource("classpath:/hadoop/hdfs-site.xml"); conf.addResource("classpath:/hadoop/mapred-site.xml"); return conf; }}
3).新建MR程序:MartrixMultiply.java
MapReduce程序
package org.conan.myhadoop.matrix;import java.io.IOException;import java.util.HashMap;import java.util.Iterator;import java.util.Map;import org.apache.hadoop.fs.Path;import org.apache.hadoop.io.IntWritable;import org.apache.hadoop.io.LongWritable;import org.apache.hadoop.io.Text;import org.apache.hadoop.mapred.JobConf;import org.apache.hadoop.mapreduce.Job;import org.apache.hadoop.mapreduce.Mapper;import org.apache.hadoop.mapreduce.Reducer;import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;import org.apache.hadoop.mapreduce.lib.input.FileSplit;import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;import org.conan.myhadoop.hdfs.HdfsDAO;public class MartrixMultiply { public static class MatrixMapper extends Mapper<LongWritable, Text, Text, Text> { private String flag;// m1 or m2 private int rowNum = 2;// 矩阵A的行数 private int colNum = 2;// 矩阵B的列数 private int rowIndexA = 1; // 矩阵A,当前在第几行 private int rowIndexB = 1; // 矩阵B,当前在第几行 @Override protected void setup(Context context) throws IOException, InterruptedException { FileSplit split = (FileSplit) context.getInputSplit(); flag = split.getPath().getName();// 判断读的数据集 } @Override public void map(LongWritable key, Text values, Context context) throws IOException, InterruptedException { String[] tokens = MainRun.DELIMITER.split(values.toString()); if (flag.equals("m1")) { for (int i = 1; i <= rowNum; i++) { Text k = new Text(rowIndexA + "," + i); for (int j = 1; j <= tokens.length; j++) { Text v = new Text("A:" + j + "," + tokens[j - 1]); context.write(k, v); System.out.println(k.toString() + " " + v.toString()); } } rowIndexA++; } else if (flag.equals("m2")) { for (int i = 1; i <= tokens.length; i++) { for (int j = 1; j <= colNum; j++) { Text k = new Text(i + "," + j); Text v = new Text("B:" + rowIndexB + "," + tokens[j - 1]); context.write(k, v); System.out.println(k.toString() + " " + v.toString()); } } rowIndexB++; } } } public static class MatrixReducer extends Reducer<Text, Text, Text, IntWritable> { @Override public void reduce(Text key, Iterable<Text> values, Context context) throws IOException, InterruptedException { Map<String, String> mapA = new HashMap<String, String>(); Map<String, String> mapB = new HashMap<String, String>(); System.out.print(key.toString() + ":"); for (Text line : values) { String val = line.toString(); System.out.print("("+val+")"); if (val.startsWith("A:")) { String[] kv = MainRun.DELIMITER.split(val.substring(2)); mapA.put(kv[0], kv[1]); // System.out.println("A:" + kv[0] + "," + kv[1]); } else if (val.startsWith("B:")) { String[] kv = MainRun.DELIMITER.split(val.substring(2)); mapB.put(kv[0], kv[1]); // System.out.println("B:" + kv[0] + "," + kv[1]); } } int result = 0; Iterator<String> iter = mapA.keySet().iterator(); while (iter.hasNext()) { String mapk = iter.next(); result += Integer.parseInt(mapA.get(mapk)) * Integer.parseInt(mapB.get(mapk)); } context.write(key, new IntWritable(result)); System.out.println(); // System.out.println("C:" + key.toString() + "," + result); } } public static void run(Map<String, String> path) throws IOException, InterruptedException, ClassNotFoundException { JobConf conf = MainRun.config(); String input = path.get("input"); String input1 = path.get("input1"); String input2 = path.get("input2"); String output = path.get("output"); HdfsDAO hdfs = new HdfsDAO(MainRun.HDFS, conf); hdfs.rmr(input); hdfs.mkdirs(input); hdfs.copyFile(path.get("m1"), input1); hdfs.copyFile(path.get("m2"), input2); Job job = new Job(conf); job.setJarByClass(MartrixMultiply.class); job.setOutputKeyClass(Text.class); job.setOutputValueClass(Text.class); job.setMapperClass(MatrixMapper.class); job.setReducerClass(MatrixReducer.class); job.setInputFormatClass(TextInputFormat.class); job.setOutputFormatClass(TextOutputFormat.class); FileInputFormat.setInputPaths(job, new Path(input1), new Path(input2));// 加载2个输入数据集 FileOutputFormat.setOutputPath(job, new Path(output)); job.waitForCompletion(true); }}
运行日志
Delete: hdfs://192.168.1.210:9000/user/hdfs/matrixCreate: hdfs://192.168.1.210:9000/user/hdfs/matrixcopy from: logfile/matrix/m1.csv to hdfs://192.168.1.210:9000/user/hdfs/matrix/m1copy from: logfile/matrix/m2.csv to hdfs://192.168.1.210:9000/user/hdfs/matrix/m22014-1-15 10:48:03 org.apache.hadoop.util.NativeCodeLoader <clinit>警告: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable2014-1-15 10:48:03 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.2014-1-15 10:48:03 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles警告: No job jar file set. User classes may not be found. See JobConf(Class) or JobConf#setJar(String).2014-1-15 10:48:03 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus信息: Total input paths to process : 22014-1-15 10:48:03 org.apache.hadoop.io.compress.snappy.LoadSnappy <clinit>警告: Snappy native library not loaded2014-1-15 10:48:04 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: Running job: job_local_00012014-1-15 10:48:04 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 10:48:04 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: io.sort.mb = 1002014-1-15 10:48:04 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: data buffer = 79691776/996147202014-1-15 10:48:04 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: record buffer = 262144/3276801,1 A:1,11,1 A:2,01,1 A:3,21,2 A:1,11,2 A:2,01,2 A:3,22,1 A:1,-12,1 A:2,32,1 A:3,12,2 A:1,-12,2 A:2,32,2 A:3,12014-1-15 10:48:04 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush信息: Starting flush of map output2014-1-15 10:48:04 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill信息: Finished spill 02014-1-15 10:48:04 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_m_000000_0 is done. And is in the process of commiting2014-1-15 10:48:05 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 0% reduce 0%2014-1-15 10:48:07 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 10:48:07 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_m_000000_0' done.2014-1-15 10:48:07 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 10:48:07 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: io.sort.mb = 1002014-1-15 10:48:07 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: data buffer = 79691776/996147202014-1-15 10:48:07 org.apache.hadoop.mapred.MapTask$MapOutputBuffer <init>信息: record buffer = 262144/3276801,1 B:1,32014-1-15 10:48:07 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush信息: Starting flush of map output1,2 B:1,12,1 B:1,32,2 B:1,11,1 B:2,21,2 B:2,12,1 B:2,22,2 B:2,11,1 B:3,11,2 B:3,02,1 B:3,12,2 B:3,02014-1-15 10:48:07 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill信息: Finished spill 02014-1-15 10:48:07 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_m_000001_0 is done. And is in the process of commiting2014-1-15 10:48:08 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 100% reduce 0%2014-1-15 10:48:10 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 10:48:10 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_m_000001_0' done.2014-1-15 10:48:10 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 10:48:10 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 10:48:10 org.apache.hadoop.mapred.Merger$MergeQueue merge信息: Merging 2 sorted segments2014-1-15 10:48:10 org.apache.hadoop.mapred.Merger$MergeQueue merge信息: Down to the last merge-pass, with 2 segments left of total size: 294 bytes2014-1-15 10:48:10 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 1,1:(B:1,3)(B:2,2)(B:3,1)(A:1,1)(A:2,0)(A:3,2)1,2:(A:1,1)(A:2,0)(A:3,2)(B:1,1)(B:2,1)(B:3,0)2,1:(B:1,3)(B:2,2)(B:3,1)(A:1,-1)(A:2,3)(A:3,1)2,2:(A:1,-1)(A:2,3)(A:3,1)(B:1,1)(B:2,1)(B:3,0)2014-1-15 10:48:10 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_r_000000_0 is done. And is in the process of commiting2014-1-15 10:48:10 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 10:48:10 org.apache.hadoop.mapred.Task commit信息: Task attempt_local_0001_r_000000_0 is allowed to commit now2014-1-15 10:48:10 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask信息: Saved output of task 'attempt_local_0001_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/matrix/output2014-1-15 10:48:13 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: reduce > reduce2014-1-15 10:48:13 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_r_000000_0' done.2014-1-15 10:48:14 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 100% reduce 100%2014-1-15 10:48:14 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: Job complete: job_local_00012014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Counters: 192014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: File Output Format Counters 2014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Bytes Written=242014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: FileSystemCounters2014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: FILE_BYTES_READ=17132014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: HDFS_BYTES_READ=752014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: FILE_BYTES_WRITTEN=1253142014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: HDFS_BYTES_WRITTEN=1142014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: File Input Format Counters 2014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Bytes Read=302014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Map-Reduce Framework2014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Map output materialized bytes=3022014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Map input records=52014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Reduce shuffle bytes=02014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Spilled Records=482014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Map output bytes=2422014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Total committed heap usage (bytes)=7642152962014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: SPLIT_RAW_BYTES=2202014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Combine input records=02014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Reduce input records=242014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Reduce input groups=42014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Combine output records=02014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Reduce output records=42014-1-15 10:48:14 org.apache.hadoop.mapred.Counters log信息: Map output records=24
4. 稀疏矩阵乘法的MapReduce计算
我们在用矩阵处理真实数据的时候,一般都是非常稀疏矩阵,为了节省存储空间,通常只会存储非0的数据。
下面我们来做一个稀疏矩阵:
- R语言的实现矩阵乘法
- 新建2个矩阵数据文件sm1.csv, sm2.csv
- 修改启动程序:MainRun.java
- 新建MR程序:SparseMartrixMultiply.java
1). R语言的实现矩阵乘法
R语言程序
> m1<-matrix(c(1,0,0,3,2,5,0,4,0,0,0,1,4,7,1,2),nrow=4,byrow=TRUE);m1 [,1] [,2] [,3] [,4][1,] 1 0 0 3[2,] 2 5 0 4[3,] 0 0 0 1[4,] 4 7 1 2> m2<-matrix(c(5,0,0,2,0,0,3,1),nrow=4,byrow=TRUE);m2 [,1] [,2][1,] 5 0[2,] 0 2[3,] 0 0[4,] 3 1> m3<-m1 %*% m2;m3 [,1] [,2][1,] 14 3[2,] 22 14[3,] 3 1[4,] 26 16
2).新建2个稀疏矩阵数据文件sm1.csv, sm2.csv
只存储非0的数据,3列存储,第一列“原矩阵行”,第二列“原矩阵列”,第三列“原矩阵值”。
sm1.csv
1,1,11,4,32,1,22,2,52,4,43,4,14,1,44,2,74,3,14,4,2
sm2.csv
1,1,52,2,24,1,34,2,1
3).修改启动程序:MainRun.java
增加SparseMartrixMultiply的启动配置
public static void main(String[] args) { sparseMartrixMultiply(); } public static void sparseMartrixMultiply() { Map<String, String> path = new HashMap<String, String>(); path.put("m1", "logfile/matrix/sm1.csv");// 本地的数据文件 path.put("m2", "logfile/matrix/sm2.csv"); path.put("input", HDFS + "/user/hdfs/matrix");// HDFS的目录 path.put("input1", HDFS + "/user/hdfs/matrix/m1"); path.put("input2", HDFS + "/user/hdfs/matrix/m2"); path.put("output", HDFS + "/user/hdfs/matrix/output"); try { SparseMartrixMultiply.run(path);// 启动程序 } catch (Exception e) { e.printStackTrace(); } System.exit(0); }
4). 新建MR程序:SparseMartrixMultiply.java
- map函数有修改,reduce函数没有变化
- 去掉判断所在行和列的变量
package org.conan.myhadoop.matrix;import java.io.IOException;import java.util.HashMap;import java.util.Iterator;import java.util.Map;import org.apache.hadoop.fs.Path;import org.apache.hadoop.io.IntWritable;import org.apache.hadoop.io.LongWritable;import org.apache.hadoop.io.Text;import org.apache.hadoop.mapred.JobConf;import org.apache.hadoop.mapreduce.Job;import org.apache.hadoop.mapreduce.Mapper;import org.apache.hadoop.mapreduce.Reducer;import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;import org.apache.hadoop.mapreduce.lib.input.FileSplit;import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;import org.conan.myhadoop.hdfs.HdfsDAO;public class SparseMartrixMultiply { public static class SparseMatrixMapper extends Mapper>LongWritable, Text, Text, Text< { private String flag;// m1 or m2 private int rowNum = 4;// 矩阵A的行数 private int colNum = 2;// 矩阵B的列数 @Override protected void setup(Context context) throws IOException, InterruptedException { FileSplit split = (FileSplit) context.getInputSplit(); flag = split.getPath().getName();// 判断读的数据集 } @Override public void map(LongWritable key, Text values, Context context) throws IOException, InterruptedException { String[] tokens = MainRun.DELIMITER.split(values.toString()); if (flag.equals("m1")) { String row = tokens[0]; String col = tokens[1]; String val = tokens[2]; for (int i = 1; i >= colNum; i++) { Text k = new Text(row + "," + i); Text v = new Text("A:" + col + "," + val); context.write(k, v); System.out.println(k.toString() + " " + v.toString()); } } else if (flag.equals("m2")) { String row = tokens[0]; String col = tokens[1]; String val = tokens[2]; for (int i = 1; i >= rowNum; i++) { Text k = new Text(i + "," + col); Text v = new Text("B:" + row + "," + val); context.write(k, v); System.out.println(k.toString() + " " + v.toString()); } } } } public static class SparseMatrixReducer extends Reducer>Text, Text, Text, IntWritable< { @Override public void reduce(Text key, Iterable>Text< values, Context context) throws IOException, InterruptedException { Map>String, String< mapA = new HashMap>String, String<(); Map>String, String< mapB = new HashMap>String, String<(); System.out.print(key.toString() + ":"); for (Text line : values) { String val = line.toString(); System.out.print("(" + val + ")"); if (val.startsWith("A:")) { String[] kv = MainRun.DELIMITER.split(val.substring(2)); mapA.put(kv[0], kv[1]); // System.out.println("A:" + kv[0] + "," + kv[1]); } else if (val.startsWith("B:")) { String[] kv = MainRun.DELIMITER.split(val.substring(2)); mapB.put(kv[0], kv[1]); // System.out.println("B:" + kv[0] + "," + kv[1]); } } int result = 0; Iterator>String< iter = mapA.keySet().iterator(); while (iter.hasNext()) { String mapk = iter.next(); String bVal = mapB.containsKey(mapk) ? mapB.get(mapk) : "0"; result += Integer.parseInt(mapA.get(mapk)) * Integer.parseInt(bVal); } context.write(key, new IntWritable(result)); System.out.println(); // System.out.println("C:" + key.toString() + "," + result); } } public static void run(Map>String, String< path) throws IOException, InterruptedException, ClassNotFoundException { JobConf conf = MainRun.config(); String input = path.get("input"); String input1 = path.get("input1"); String input2 = path.get("input2"); String output = path.get("output"); HdfsDAO hdfs = new HdfsDAO(MainRun.HDFS, conf); hdfs.rmr(input); hdfs.mkdirs(input); hdfs.copyFile(path.get("m1"), input1); hdfs.copyFile(path.get("m2"), input2); Job job = new Job(conf); job.setJarByClass(MartrixMultiply.class); job.setOutputKeyClass(Text.class); job.setOutputValueClass(Text.class); job.setMapperClass(SparseMatrixMapper.class); job.setReducerClass(SparseMatrixReducer.class); job.setInputFormatClass(TextInputFormat.class); job.setOutputFormatClass(TextOutputFormat.class); FileInputFormat.setInputPaths(job, new Path(input1), new Path(input2));// 加载2个输入数据集 FileOutputFormat.setOutputPath(job, new Path(output)); job.waitForCompletion(true); }}
运行输出:
Delete: hdfs://192.168.1.210:9000/user/hdfs/matrixCreate: hdfs://192.168.1.210:9000/user/hdfs/matrixcopy from: logfile/matrix/sm1.csv to hdfs://192.168.1.210:9000/user/hdfs/matrix/m1copy from: logfile/matrix/sm2.csv to hdfs://192.168.1.210:9000/user/hdfs/matrix/m22014-1-15 11:57:31 org.apache.hadoop.util.NativeCodeLoader >clinit<警告: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable2014-1-15 11:57:31 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles警告: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.2014-1-15 11:57:31 org.apache.hadoop.mapred.JobClient copyAndConfigureFiles警告: No job jar file set. User classes may not be found. See JobConf(Class) or JobConf#setJar(String).2014-1-15 11:57:31 org.apache.hadoop.mapreduce.lib.input.FileInputFormat listStatus信息: Total input paths to process : 22014-1-15 11:57:31 org.apache.hadoop.io.compress.snappy.LoadSnappy >clinit<警告: Snappy native library not loaded2014-1-15 11:57:31 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: Running job: job_local_00012014-1-15 11:57:31 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 11:57:31 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: io.sort.mb = 1002014-1-15 11:57:31 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: data buffer = 79691776/996147202014-1-15 11:57:31 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: record buffer = 262144/3276801,1 A:1,11,2 A:1,11,1 A:4,31,2 A:4,32,1 A:1,22,2 A:1,22,1 A:2,52,2 A:2,52,1 A:4,42,2 A:4,43,1 A:4,13,2 A:4,14,1 A:1,44,2 A:1,44,1 A:2,74,2 A:2,74,1 A:3,14,2 A:3,14,1 A:4,24,2 A:4,22014-1-15 11:57:31 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush信息: Starting flush of map output2014-1-15 11:57:31 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill信息: Finished spill 02014-1-15 11:57:31 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_m_000000_0 is done. And is in the process of commiting2014-1-15 11:57:32 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 0% reduce 0%2014-1-15 11:57:34 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 11:57:34 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_m_000000_0' done.2014-1-15 11:57:34 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 11:57:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: io.sort.mb = 1002014-1-15 11:57:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: data buffer = 79691776/996147202014-1-15 11:57:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer >init<信息: record buffer = 262144/3276801,1 B:1,52,1 B:1,53,1 B:1,54,1 B:1,52014-1-15 11:57:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer flush信息: Starting flush of map output1,2 B:2,22,2 B:2,23,2 B:2,24,2 B:2,21,1 B:4,32,1 B:4,33,1 B:4,34,1 B:4,31,2 B:4,12,2 B:4,13,2 B:4,14,2 B:4,12014-1-15 11:57:34 org.apache.hadoop.mapred.MapTask$MapOutputBuffer sortAndSpill信息: Finished spill 02014-1-15 11:57:34 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_m_000001_0 is done. And is in the process of commiting2014-1-15 11:57:35 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 100% reduce 0%2014-1-15 11:57:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 11:57:37 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_m_000001_0' done.2014-1-15 11:57:37 org.apache.hadoop.mapred.Task initialize信息: Using ResourceCalculatorPlugin : null2014-1-15 11:57:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 11:57:37 org.apache.hadoop.mapred.Merger$MergeQueue merge信息: Merging 2 sorted segments2014-1-15 11:57:37 org.apache.hadoop.mapred.Merger$MergeQueue merge信息: Down to the last merge-pass, with 2 segments left of total size: 436 bytes2014-1-15 11:57:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 1,1:(B:1,5)(B:4,3)(A:1,1)(A:4,3)1,2:(A:1,1)(A:4,3)(B:2,2)(B:4,1)2,1:(B:1,5)(B:4,3)(A:1,2)(A:2,5)(A:4,4)2,2:(A:1,2)(A:2,5)(A:4,4)(B:4,1)(B:2,2)3,1:(B:1,5)(B:4,3)(A:4,1)3,2:(A:4,1)(B:2,2)(B:4,1)4,1:(B:4,3)(B:1,5)(A:1,4)(A:2,7)(A:3,1)(A:4,2)4,2:(A:1,4)(A:2,7)(A:3,1)(A:4,2)(B:2,2)(B:4,1)2014-1-15 11:57:37 org.apache.hadoop.mapred.Task done信息: Task:attempt_local_0001_r_000000_0 is done. And is in the process of commiting2014-1-15 11:57:37 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: 2014-1-15 11:57:37 org.apache.hadoop.mapred.Task commit信息: Task attempt_local_0001_r_000000_0 is allowed to commit now2014-1-15 11:57:37 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter commitTask信息: Saved output of task 'attempt_local_0001_r_000000_0' to hdfs://192.168.1.210:9000/user/hdfs/matrix/output2014-1-15 11:57:40 org.apache.hadoop.mapred.LocalJobRunner$Job statusUpdate信息: reduce < reduce2014-1-15 11:57:40 org.apache.hadoop.mapred.Task sendDone信息: Task 'attempt_local_0001_r_000000_0' done.2014-1-15 11:57:41 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: map 100% reduce 100%2014-1-15 11:57:41 org.apache.hadoop.mapred.JobClient monitorAndPrintJob信息: Job complete: job_local_00012014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Counters: 192014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: File Output Format Counters 2014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Bytes Written=532014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: FileSystemCounters2014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: FILE_BYTES_READ=25032014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: HDFS_BYTES_READ=2662014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: FILE_BYTES_WRITTEN=1262742014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: HDFS_BYTES_WRITTEN=3472014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: File Input Format Counters 2014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Bytes Read=982014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Map-Reduce Framework2014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Map output materialized bytes=4442014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Map input records=142014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Reduce shuffle bytes=02014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Spilled Records=722014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Map output bytes=3602014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Total committed heap usage (bytes)=7642152962014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: SPLIT_RAW_BYTES=2202014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Combine input records=02014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Reduce input records=362014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Reduce input groups=82014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Combine output records=02014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Reduce output records=82014-1-15 11:57:41 org.apache.hadoop.mapred.Counters log信息: Map output records=36
程序源代码,已上传到github:
https://github.com/bsspirit/maven_hadoop_template/tree/master/src/main/java/org/conan/myhadoop/matrix
这样就用MapReduce的程序,实现了矩阵的乘法!有了矩阵计算的基础,接下来,我们就可以做更多的事情了!
- 用MapReduce实现矩阵乘法
- 用MapReduce实现矩阵乘法
- 用MapReduce实现矩阵乘法
- MapReduce实现矩阵乘法
- Mapreduce实现矩阵乘法
- MapReduce实现矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- MapReduce实现大矩阵乘法
- 另一种MapReduce实现矩阵乘法
- Hadoop MapReduce实现矩阵的乘法
- 矩阵乘法的MapReduce程序设计与实现
- android GridView实现图库预览图,多选模式下右上角打勾
- c#和JAVA之间使用RSA加解密交互1
- TabHost中的Activity调用overridePendingTransition()方法无效
- 格林威尔北向接口在电信中的应用
- shell编程和python有什么不同
- 用MapReduce实现矩阵乘法
- Java.lang.UnsatisfiedLinkError
- django模板放置的位置
- SHELL学习笔记3
- C# Java间进行RSA加密解密交互(二)
- 有关socket AF_UNIX的一点心得
- 微信企业号开发之获取AccessToken
- Activity切换效果(overridePendingTransition)
- 程序设计基本概念——运算符