Spark Streaming实时读取Kafka
来源:互联网 发布:商务双肩包 知乎 编辑:程序博客网 时间:2024/06/06 00:53
使用Spark Streaming读取Kafka时,要引用如下包
<dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-streaming_2.10</artifactId> <version>1.6.1</version></dependency> <dependency> <groupId>org.apache.kafka</groupId> <artifactId>kafka-clients</artifactId> <version>0.10.0.0</version></dependency><dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-streaming-kafka_2.10</artifactId> <version>1.6.1</version></dependency>
以下是使用createDirectStream读取Kafka,该方法比createStream方法读取速度快很多!
import org.apache.spark.SparkConfimport org.apache.spark.SparkContextimport org.apache.spark.SparkContext._import org.apache.spark.streaming.{Seconds, StreamingContext}import org.apache.spark.streaming.StreamingContext._import org.apache.spark.streaming.kafka._import kafka.serializer.StringDecoder import kafka.serializer.DefaultDecoderimport org.apache.spark.serializer.KryoSerializerobject DirectKafka{ def main(args: Array[String]) { val brokers = "192.168.2.5:9092" val topics = "topic-name" val sparkConf = new SparkConf().set("spark.driver.cores", "8").setAppName("DirectKafka") val ssc = new StreamingContext(sparkConf, Seconds(1)) val topicsSet = topics.split(",").toSet val kafkaParams = Map[String, String]( "metadata.broker.list" -> brokers ) val messages = KafkaUtils.createDirectStream[String, Array[Byte], StringDecoder, DefaultDecoder]( ssc, kafkaParams, topicsSet).map(_._2) messages.foreachRDD( x => println("each RDD count : " + x.count()) ) ssc.start() ssc.awaitTermination() }}
0 0
- Spark Streaming实时读取Kafka
- spark streaming 读取kafka 配置
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- Spark Streaming+kafka订单实时统计实现
- Kafka+Spark Streaming+Redis实时系统实践
- Kafka+Spark Streaming+Redis实时系统实践
- kafka + spark streaming 实时读取计算 nginx 日志,存储结果到 mongodb/mysql
- spark streaming 读取kafka的offset
- spark streaming 读取kafka数据问题
- spark streaming读取kafka数据,记录offset
- spark-streaming 读取kafka的方式
- Spark Streaming 读取Kafka数据写入Elasticsearch
- 浅析php变量存储结构及弱类型实现机制
- Head first PHP&MySQL 学习笔记(一)
- HTML4和HTML5差距
- SMS短信的PDU编码规则
- ubuntu下Android Studio 提交代码至Github
- Spark Streaming实时读取Kafka
- OpenGL中的3D文字绘制
- NSURLSession和NSURLConnection
- Flask Web 开发 用户资料
- 在限定期限内索取一套专属于您的免费Delphi 10.1 Berlin Starter版本软件开发工具
- Linux-- 安装软件(centos下)
- Spring AOP + Redis缓存数据库查询
- MySQL:讨人喜欢的 MySQL replace into 用法(insert into 的增强版)
- Service学习之路(02)之Service启动和停止Service