Spark函数讲解:collectAsMap
来源:互联网 发布:三体 死神永生 知乎 编辑:程序博客网 时间:2024/05/19 13:55
功能和collect函数类似。该函数用于Pair RDD,最终返回Map类型的结果。官方文档说明:
Return the key-value pairs in this RDD to the master as a Map.
Warning: this doesn't return a multimap (so if you have multiple values to the same key, only one value per key is preserved in the map returned)
Return the key-value pairs in this RDD to the master as a Map.
Warning: this doesn't return a multimap (so if you have multiple values to the same key, only one value per key is preserved in the map returned)
函数原型
def collectAsMap(): Map[K, V]
实例
scala> val data = sc.parallelize(List((1, "www"), (1, "iteblog"), (1, "com"), (2, "bbs"), (2, "iteblog"), (2, "com"), (3, "good")))data: org.apache.spark.rdd.RDD[(Int, String)] = ParallelCollectionRDD[26] at parallelize at <console>:12scala> data.collectAsMapres28: scala.collection.Map[Int,String] = Map(2 -> com, 1 -> com, 3 -> good)从结果我们可以看出,如果RDD中同一个Key中存在多个Value,那么后面的Value将会把前面的Value覆盖,最终得到的结果就是Key唯一,而且对应一个Value。
0 0
- Spark函数讲解:collectAsMap
- Spark函数讲解:collectAsMap
- Spark函数讲解:aggregate
- Spark函数讲解:cogroup
- Spark函数讲解:coalesce
- Spark函数讲解:checkpoint
- Spark函数讲解:cartesian
- Spark函数讲解:cache
- Spark函数讲解:aggregateByKey
- Spark函数讲解:collect
- Spark函数讲解:combineByKey
- Spark函数讲解:coalesce
- Spark函数讲解:aggregateByKey
- spark函数讲解:cogroup
- spark函数讲解:aggregate
- Spark算子:RDDAction操作–first/count/reduce/collect/collectAsMap
- Spark函数讲解序列文章
- Spark常用函数讲解--键值RDD转换
- EventBus3.0使用详解
- Break and Continue
- vs2010与oracle数据库连接
- Unity3D 游戏序列化及发布
- HTML-通过点击网页上的文字弹出QQ添加好友页面
- Spark函数讲解:collectAsMap
- 杭电OJ 1010, 从非递归的DFS 到 剪枝
- 百度之星2016初赛第一场(Astar Round 2A)
- vim折叠
- 感恩父母,不要等到子欲养而亲不待
- 共用体union
- android-项目目录的详细介绍
- grivade刷新success
- 大小端模式