【spark】spark+kafka
来源:互联网 发布:mac apache 无法访问 编辑:程序博客网 时间:2024/06/03 23:38
:启动kafka
MobaXterm_Personal_8.5.exe
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/zookeeper-server-start.bat D:/Develop/kafka_2.10-0.8.2.1/config/zookeeper.properties
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-server-start.bat D:/Develop/kafka_2.10-0.8.2.1/config/server.properties
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-topics.bat --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic test
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-topics.bat --list --zookeeper localhost:2181
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-console-producer.bat --broker-list localhost:9092 --topic test
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-console-consumer.bat --zookeeper localhost:2181 --topic test --from-beginning
MobaXterm_Personal_8.5.exe
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/zookeeper-server-start.bat D:/Develop/kafka_2.10-0.8.2.1/config/zookeeper.properties
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-server-start.bat D:/Develop/kafka_2.10-0.8.2.1/config/server.properties
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-topics.bat --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic test
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-topics.bat --list --zookeeper localhost:2181
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-console-producer.bat --broker-list localhost:9092 --topic test
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-console-consumer.bat --zookeeper localhost:2181 --topic test --from-beginning
D:/Develop/kafka_2.10-0.8.2.1/bin/windows/kafka-topics.bat --describe --zookeeper localhost:2181 --topic test
:连接kafka
package com.baibaiw5.ml.antispamimport org.apache.spark.SparkConfimport org.apache.spark.storage.StorageLevelimport org.apache.spark.streaming.kafka.KafkaUtilsimport org.apache.spark.streaming.{ Seconds, StreamingContext }import org.apache.spark.Loggingobject EmailSpamStreamingApp extends Logging { def main(args: Array[String]) { logInfo("start EmailSpamStreamingApp") val zk = "localhost:2181" val group = "my-group" val topics = "test" val sparkConf = new SparkConf().setMaster("local[2]").setAppName("EmailSpamStreamingApp") val sc = new StreamingContext(sparkConf, Seconds(5)) //should be lease than cup cores and local[2] val numThreads = 1 val topicpMap = topics.split(",").map((_, numThreads.toInt)).toMap logInfo("create kafka consumer") val lines = KafkaUtils.createStream(sc, zk, group, topicpMap, StorageLevel.MEMORY_AND_DISK_SER).map(_._2) val words = lines.flatMap(_.split(" ")) val pairs = words.map(word => (word, 1)) val wordCounts = pairs.reduceByKey(_ + _) wordCounts.print() logInfo("start streamnig") sc.start() sc.awaitTermination() }}
0 0
- 【spark】spark+kafka
- spark-kafka
- Spark Streaming整合Kafka
- Spark Streaming + kafka
- Kafka+Spark Streaming集成
- kafka + spark streaming 架构
- spark kafka mysql
- spark kafka hbase
- Spark+kafka+SparkStreaming实例
- spark streaming+kafka
- Spark + Kafka 集成 指南
- Spark Streaming + Kafka
- kafka-spark streaming-hbase
- spark整合kafka案例
- spark streaming 整合kafka
- spark streaming + kafka
- flume kafka spark streaming
- Kafka->Spark Streaming->mongodb
- ecshop ecs_template 解释 作用,数据字段 首页自定义分类显示,美乐乐
- vs2012 opencv 模块计算机类新x86与目标计算机类x64冲突
- 不敢死队问题
- 测试
- Leetcode 100. Same Tree
- 【spark】spark+kafka
- tomcat的目录结构
- hdu 5675 ztr loves math(数学)
- ArrayList LinkedList Vector的区别
- kafka丢数据和数据重复
- POJ Anniversary Cake (dfs,神剪)
- Android arcgis 坐标转经纬度
- Common Subsequence
- 电池分流器之电路