java kafka消息的发送与接收
消息队列在java EE级开发是很常用到的工具之一,在众多消息队列当中,active mq与kafka相对比较受开发者的喜爱,那么kafka是怎样实现消息的发送与接收呢?这里我们的消息通过一个实体类对象来进行封装,前提是你服务器上已经搭建好kafka环境,整个代码如下:
1、kafka工具类,实现了生产者消费者消息的封装,以及发送消息到kafka上和从kafka上获取消息。
1、kafka工具类,实现了生产者消费者消息的封装,以及发送消息到kafka上和从kafka上获取消息。
- package com.starit.ipran.kafka;
- import java.util.ArrayList;
- import java.util.Arrays;
- import java.util.Date;
- import java.util.HashMap;
- import java.util.List;
- import java.util.Map;
- import java.util.Properties;
- import org.apache.kafka.clients.consumer.Consumer;
- import org.apache.kafka.clients.consumer.ConsumerRecord;
- import org.apache.kafka.clients.consumer.ConsumerRecords;
- import org.apache.kafka.clients.consumer.KafkaConsumer;
- import org.apache.kafka.clients.producer.KafkaProducer;
- import org.apache.kafka.clients.producer.Producer;
- import org.apache.kafka.clients.producer.ProducerRecord;
- import org.slf4j.Logger;
- import org.slf4j.LoggerFactory;
- import com.alibaba.fastjson.JSON;
- import com.alibaba.fastjson.JSONObject;
- import com.starit.ipran.load.Constants;
- import com.starit.ipran.model.IpranAlarm;
- public class KafkaUtils {
- private final static Logger LOGGER = LoggerFactory.getLogger(KafkaUtils.class);
- private static Producer<String, String> producer;
- private static Consumer<String, String> consumer;
- private KafkaUtils() {
- }
- /**
- * 生产者,注意kafka生产者不能够从代码上生成主题,只有在服务器上用命令生成
- */
- static {
- Properties props = new Properties();
- props.put("bootstrap.servers", Constants.BOOTSTRAP_SERVERS);//服务器ip:端口号,集群用逗号分隔
- props.put("acks", "all");
- props.put("retries", 0);
- props.put("batch.size", 16384);
- props.put("linger.ms", 1);
- props.put("buffer.memory", 33554432);
- props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
- props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
- producer = new KafkaProducer<>(props);
- }
- /**
- * 消费者
- */
- static {
- Properties props = new Properties();
- props.put("bootstrap.servers", Constants.BOOTSTRAP_SERVERS);//服务器ip:端口号,集群用逗号分隔
- props.put("group.id", "test");
- props.put("enable.auto.commit", "true");
- props.put("auto.commit.interval.ms", "1000");
- props.put("session.timeout.ms", "30000");
- props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
- props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
- consumer = new KafkaConsumer<>(props);
- consumer.subscribe(Arrays.asList(Constants.TOPIC_NAME));
- }
- /**
- * 发送对象消息 至kafka上,调用json转化为json字符串,应为kafka存储的是String。
- * @param msg
- */
- public static void sendMsgToKafka(IpranAlarm msg) {
- producer.send(new ProducerRecord<String, String>(Constants.TOPIC_NAME, String.valueOf(new Date().getTime()),
- JSON.toJSONString(msg)));
- }
- /**
- * 从kafka上接收对象消息,将json字符串转化为对象,便于获取消息的时候可以使用get方法获取。
- */
- public static void getMsgFromKafka(){
- while(true){
- ConsumerRecords<String, String> records = KafkaUtils.getKafkaConsumer().poll(100);
- if (records.count() > 0) {
- for (ConsumerRecord<String, String> record : records) {
- JSONObject jsonAlarmMsg = JSON.parseObject(record.value());
- IpranAlarm alarmMsg = JSONObject.toJavaObject(jsonAlarmMsg, IpranAlarm.class);
- LOGGER.info("从kafka接收到的消息是:" + alarmMsg.toString());
- }
- }
- }
- }
- public static Consumer<String, String> getKafkaConsumer() {
- return consumer;
- }
- public static void closeKafkaProducer() {
- producer.close();
- }
- public static void closeKafkaConsumer() {
- consumer.close();
- }
- }