kafka生产者、消费者java示例
来源:互联网 发布:js json数组添加元素 编辑:程序博客网 时间:2024/04/26 22:33
1. 生产者
import java.util.Properties; import kafka.javaapi.producer.Producer; import kafka.producer.KeyedMessage; import kafka.producer.ProducerConfig; public class MyProducer { public static void main(String[] args) { Properties props = new Properties(); props.setProperty("metadata.broker.list","localhost:9092"); props.setProperty("serializer.class","kafka.serializer.StringEncoder"); props.put("request.required.acks","1"); ProducerConfig config = new ProducerConfig(props); //创建生产这对象 Producer<String, String> producer = new Producer<String, String>(config); //生成消息 KeyedMessage<String, String> data = new KeyedMessage<String, String>("mykafka","test-kafka"); try { int i =1; while(i < 100){ //发送消息 producer.send(data); } } catch (Exception e) { e.printStackTrace(); } producer.close(); } }
2. 消费者
import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.Properties; import kafka.consumer.ConsumerConfig; import kafka.consumer.ConsumerIterator; import kafka.consumer.KafkaStream; import kafka.javaapi.consumer.ConsumerConnector; public class MyConsumer extends Thread{ //消费者连接 private final ConsumerConnector consumer; //要消费的话题 private final String topic; public MyConsumer(String topic) { consumer =kafka.consumer.Consumer .createJavaConsumerConnector(createConsumerConfig()); this.topic =topic; } //配置相关信息 private static ConsumerConfig createConsumerConfig() { Properties props = new Properties(); // props.put("zookeeper.connect","localhost:2181,10.XX.XX.XX:2181,10.XX.XX.XX:2181"); //配置要连接的zookeeper地址与端口 //The ‘zookeeper.connect’ string identifies where to find once instance of Zookeeper in your cluster. //Kafka uses ZooKeeper to store offsets of messages consumed for a specific topic and partition by this Consumer Group props.put("zookeeper.connect","localhost:2181"); //配置zookeeper的组id (The ‘group.id’ string defines the Consumer Group this process is consuming on behalf of.) props.put("group.id", "0"); //配置zookeeper连接超时间隔 //The ‘zookeeper.session.timeout.ms’ is how many milliseconds Kafka will wait for //ZooKeeper to respond to a request (read or write) before giving up and continuing to consume messages. props.put("zookeeper.session.timeout.ms","10000"); //The ‘zookeeper.sync.time.ms’ is the number of milliseconds a ZooKeeper ‘follower’ can be behind the master before an error occurs. props.put("zookeeper.sync.time.ms", "200"); //The ‘auto.commit.interval.ms’ setting is how often updates to the consumed offsets are written to ZooKeeper. //Note that since the commit frequency is time based instead of # of messages consumed, if an error occurs between updates to ZooKeeper on restart you will get replayed messages. props.put("auto.commit.interval.ms", "1000"); return new ConsumerConfig(props); } public void run(){ Map<String,Integer> topickMap = new HashMap<String, Integer>(); topickMap.put(topic, 1); Map<String, List<KafkaStream<byte[],byte[]>>> streamMap =consumer.createMessageStreams(topickMap); KafkaStream<byte[],byte[]>stream = streamMap.get(topic).get(0); ConsumerIterator<byte[],byte[]> it =stream.iterator(); System.out.println("*********Results********"); while(true){ if(it.hasNext()){ //打印得到的消息 System.err.println(Thread.currentThread()+" get data:" +new String(it.next().message())); } try { Thread.sleep(1000); } catch (InterruptedException e) { e.printStackTrace(); } } } public static void main(String[] args) { MyConsumer consumerThread = new MyConsumer("mykafka"); consumerThread.start(); } }
3. 消费者的线程执行器实现
首先建立一个处理消息的类Consumer
import kafka.consumer.ConsumerIterator;import kafka.consumer.KafkaStream;public class Consumer implements Runnable { private KafkaStream stream; private int threadNumber; public Consumer(KafkaStream a_stream, int a_threadNumber) { threadNumber = a_threadNumber; stream = a_stream; } public void run() { ConsumerIterator<byte[], byte[]> it = stream.iterator(); while (it.hasNext()) System.out.println("Thread " + threadNumber + ": " + new String(it.next().message())); System.out.println("Shutting down Thread: " + threadNumber); }}
其次实现多线程的调用
import kafka.consumer.ConsumerConfig;import kafka.consumer.KafkaStream;import kafka.javaapi.consumer.ConsumerConnector;import java.util.concurrent.*; import java.util.HashMap;import java.util.List;import java.util.Map;import java.util.Properties;import java.util.concurrent.ExecutorService;import java.util.concurrent.Executors; public class ConsumerGroup { private final ConsumerConnector consumer; private final String topic; private ExecutorService executor; public ConsumerGroup(String a_zookeeper, String a_groupId, String a_topic) { consumer = kafka.consumer.Consumer.createJavaConsumerConnector( createConsumerConfig(a_zookeeper, a_groupId)); this.topic = a_topic; } public void shutdown() { if (consumer != null) consumer.shutdown(); if (executor != null) executor.shutdown(); try { if (!executor.awaitTermination(5000, TimeUnit.MILLISECONDS)) { System.out.println("Timed out waiting for consumer threads to shut down, exiting uncleanly"); } } catch (InterruptedException e) { System.out.println("Interrupted during shutdown, exiting uncleanly"); } } public void run(int a_numThreads) { Map<String, Integer> topicCountMap = new HashMap<String, Integer>(); topicCountMap.put(topic, new Integer(a_numThreads)); Map<String, List<KafkaStream<byte[], byte[]>>> consumerMap = consumer.createMessageStreams(topicCountMap); List<KafkaStream<byte[], byte[]>> streams = consumerMap.get(topic); // now launch all the threads // executor = Executors.newFixedThreadPool(a_numThreads); // now create an object to consume the messages // int threadNumber = 0; for (final KafkaStream stream : streams) { executor.submit(new Consumer(stream, threadNumber)); threadNumber++; } } private static ConsumerConfig createConsumerConfig(String a_zookeeper, String a_groupId) { Properties props = new Properties(); props.put("zookeeper.connect", a_zookeeper); props.put("group.id", a_groupId); props.put("zookeeper.session.timeout.ms", "400"); props.put("zookeeper.sync.time.ms", "200"); props.put("auto.commit.interval.ms", "1000"); return new ConsumerConfig(props); } public static void main(String[] args) { String zooKeeper = "localhost:2181"; String groupId = "0"; String topic = "mykafka"; int threads = 2; //启动的线程数 ConsumerGroup group = new ConsumerGroup(zooKeeper, groupId, topic); group.run(threads); try { Thread.sleep(10000); } catch (InterruptedException ie) { } group.shutdown(); }}
0 0
- kafka生产者、消费者java示例
- kafka生产者与消费者java代码示例
- Kafka生产者消费者java示例(包含Avro序列化)
- kafka本地java示例生产者与消费者,非ZK版
- Java生产者消费者示例
- kafka java 生产者消费者demo
- java实现Kafka生产者示例
- 使用Java写kafka生产者消费者
- java 实现kafka消息生产者和消费者
- Kafka 生产者和消费者 demo (java&scala)
- kafka 生产者消费者配置
- kafka生产者消费者
- kafka C++ 生产者 消费者
- Kafka消费者生产者实例
- kafka生产者消费者
- kafka集群搭建和使用Java写kafka生产者消费者
- kafka集群搭建和使用Java写kafka生产者消费者
- kafka集群搭建和使用Java写kafka生产者消费者
- SpringMVC的JSON实现
- web前端之google地图的封装 var mapConfigUrl = "TzxMap/TzxMapBaseInfoJson.txt"; var mapType = "goo
- 追踪oracle执行sql情况
- Java设计模式-单例模式
- hdu 3336
- kafka生产者、消费者java示例
- [C#]结构体和类的区别
- jquery中对于select框的取值和赋值
- 卫监接口项目一些零碎
- 【回溯法】游戏竞赛
- Swift:Call can throw, but it is not marked with 'try' and the error is not handled
- 深入浅出UML类图
- 关于limit_req和limit_conn的区别
- Android中利用Camera与Matrix实现3D效果详解