2019独角兽企业重金招聘Python工程师标准>>>
上一篇文章我们讲到了Kafka的工作原理和如何使用Kafka的代码示例,这里我们开始讲解Kafka的实战,在实际的应用中我们如何使用kafka的。下面将介绍前台的操作日志定时推送到kafka,然后通过kafka将消息日志进行保存,方便大数据的统计分析形成运营报表。
我们先看看工程的目录结构:
kafka的版本是:
org.apache.kafka
kafka_2.10
0.10.2.0
下面我们依次看下代码实现:
错误码字典类CodeConstant.java
public class CodeConstant {
// 参数为空
public static int NULL_ERROR = -1;
// 请求参数错误
public static int PARAM_ERROR = -2;
// token错误
public static int TOKEN_ERROR = -3;
}
返回信息实体类JsonMsg.java
public class JsonMsg {
private int code;
private String message;
public int getCode() {
return code;
}
public void setCode(int code) {
this.code = code;
}
public String getMessage() {
return message;
}
public void setMessage(String message) {
this.message = message;
}
}
kafka消息实体类Message.java
/**
* kafka消息实体类
* @author fuyuwei
* 2017年6月10日 下午10:57:17
*/
public class Message implements Serializable {
private static final long serialVersionUID = -6170235919490993626L;
/**
* 消息主键
*/
protected String messageId;
/**
* 回复消息对应的源消息主键
*/
protected String sourceMessageId;
/**
* 发送消息相关信息
*/
protected String sender;
/**
* 消息体
*/
protected byte[] messageBody;
/**
* 消息创建时间
*/
protected long createTime;
public Message(byte[] messageBody){
this.sender = getIps();
createMessageId();
this.messageBody = messageBody;
this.createTime = System.currentTimeMillis();
}
public String getIps(){
try {
return InetAddress.getLocalHost().getHostAddress();
} catch (UnknownHostException e) {
e.printStackTrace();
}
return "";
}
/**
* 消息转为在消息中间件传输的内容
* @return
* @throws BusinessException
*/
public String toJSONString() throws BusinessException {
createMessageId();
try {
return JsonUtil.toJSon(this);
} catch (BusinessException e) {
throw e;
}
}
/**
* 接收到的消息转为实体对象
* @param content 消息内容
* @return 消息实体
* @throws BusinessException
*/
public Message toMessage(String content) throws BusinessException{
return JsonUtil.readValue(content, Message.class);
}
public String toString(){
String date =null;
try {
SimpleDateFormat sdf=new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
date = sdf.format(new Date(createTime));
} catch (Exception e) {
}
StringBuffer sb = new StringBuffer();
sb.append("messageId:"+this.messageId+"\r\n").append("sourceMessageId:"+this.messageId+"\r\n")
.append("sender:"+sender+"\r\n").append("messageBody"+messageBody+"\r\n")
.append("createTime="+date+"\r\n");
return sb.toString();
}
public String getMessageId() {
return messageId;
}
private void createMessageId() {
this.messageId = sender+createUUID();
}
private String createUUID(){
String id = UUID.randomUUID().toString();
return id.substring(0,8)+id.substring(9,13)+id.substring(14,18)+id.substring(19,23)+id.substring(24); //去掉“-”符号
}
public String getSender() {
return sender;
}
public void setSender(String sender) {
this.sender = sender;
}
public long getCreateTime() {
return createTime;
}
public void setCreateTime(long createTime) {
this.createTime = createTime;
}
public String getSourceMessageId() {
return sourceMessageId;
}
public void setSourceMessageId(String sourceMessageId) {
this.sourceMessageId = sourceMessageId;
}
public byte[] getMessageBody() {
return messageBody;
}
public void setMessageBody(byte[] messageBody) {
this.messageBody = messageBody;
}
public void setMessageId(String messageId) {
this.messageId = messageId;
}
}
Http访问类KafkaLogController.java
/**
* kafka log接收器
* @author fuyuwei
* 2017年6月10日 下午8:00:07
*/
@Controller
public class KafkaLogController {
private static MessageProducer producer = MessageProducer.getInstance();
/**
* 接收前台传来的日志字符串
* 既然采用Http协议请求,务必考虑传输的安全性,添加了请求的参数拦截校验
* @author fuyuwei
* 2017年6月10日 下午8:01:36
* @param req
* @param resp
* @throws Throwable
*/
@RequestMapping(value = "/kafka/log/receiveLog.do", method= RequestMethod.POST)
public void receiveLog(HttpServletRequest req, HttpServletResponse resp) throws Throwable{
ServletInputStream is = req.getInputStream();
byte[] bytes = readStream(is);
if(bytes == null || bytes.length == 0){
JsonMsg msg = new JsonMsg();
msg.setCode(CodeConstant.NULL_ERROR);
msg.setMessage("the request data is null");
// 不设置缓存
RespUtil.responJson(resp, msg, 0);
return;
}
Message message = new Message(bytes);
producer.sendMessage("appLog", message);
BizLogger.info("receiveLog","appLog",message.getMessageId());
}
/**
* 把日志字符串转换为字节流数组
* @author fuyuwei
* 2017年6月10日 下午8:05:20
* @param inStream
* @return
*/
public static byte[] readStream(InputStream inStream){
ByteArrayOutputStream outStream = new ByteArrayOutputStream();
BufferedInputStream inputStream = new BufferedInputStream(inStream);
try {
byte[] buffer = new byte[1024];
int len = -1;
while((len = inputStream.read(buffer)) != -1){
outStream.write(buffer,0,len);
}
return outStream.toByteArray();
} catch (IOException e) {
BizLogger.error(e, "inputStream.read failure...");
}
return null;
}
}
Spring启动加载类InitMessageConsumer.java
public class InitMessageConsumer implements InitializingBean, DisposableBean {
public MessageConsumer consumer;
@Override
public void destroy() throws Exception {
}
@Override
public void afterPropertiesSet() throws Exception {
}
public void initMethod() {
BizLogger.info("init MessageReceiver start");
consumer = new MessageConsumer("appLog", 2,"app-group", new MessageConsumerExecutor());
try {
consumer.receiveMessage();
} catch (Exception e) {
BizLogger.error(e, "InitAndDestroySeqBean initMethod");
}
BizLogger.info("MessageReceiver init finish!");
}
public void destroyMethod() {
if (null != consumer) {
consumer.close();
}
}
}
拦截器AccessInteceptor.java
public class AccessInteceptor implements HandlerInterceptor {
@Override
public void afterCompletion(HttpServletRequest req,
HttpServletResponse res, Object o, Exception e) throws Exception {
}
@Override
public void postHandle(HttpServletRequest req, HttpServletResponse res,
Object o, ModelAndView m) throws Exception {
}
@Override
public boolean preHandle(HttpServletRequest req, HttpServletResponse res,
Object o) throws Exception {
String flagImmei = req.getHeader("flagImmei");
String tk = req.getHeader("token");
if(flagImmei.length() > 40){
JsonMsg msg = new JsonMsg();
msg.setCode(CodeConstant.PARAM_ERROR);
msg.setMessage("the request data is null");
// 不设置缓存
RespUtil.responJson(res, msg, 0);
return false;
}
if(!AppAESUtil.check(tk, flagImmei)){
JsonMsg msg = new JsonMsg();
msg.setCode(CodeConstant.TOKEN_ERROR);
msg.setMessage("the token is error");
RespUtil.responJson(res, msg, 0);
return false;
}
return true;
}
}
消息生产者MessageProducer.java
public class MessageProducer implements MessageService {
private Producer producer;
private static MessageProducer instance = null;
/**
* 初始化生产者
*/
private MessageProducer() {
try {
Properties properties = new Properties();
properties.load(new ClassPathResource("producer.properties").getInputStream());
producer = new KafkaProducer<>(properties);
} catch (IOException e) {
BizLogger.error(e, "load producer file fail!");
}
}
/**
* 单例模式
* @author fuyuwei
* 2017年6月10日 下午8:44:05
* @return
*/
public static synchronized MessageProducer getInstance() {
if(instance == null){
synchronized(MessageProducer.class){
if(instance == null){
instance = new MessageProducer();
}
}
}
return instance;
}
/**
* 发送消息
*/
public boolean sendMessage(String topic, Message message) throws Exception {
Collection messages = new ArrayList();
messages.add(message);
return sendMessage(topic, messages);
}
/**
* 批量发送消息
*/
public boolean sendMessage(String topic, Collection messages) throws Exception {
if (messages == null || messages.isEmpty()) {
return false;
}
for (Message message : messages) {
ProducerRecord km = new ProducerRecord(topic, message.getMessageId(),
message.getMessageBody());
producer.send(km);
}
return true;
}
/**
* 关闭发送客户端
*/
public void close() {
producer.close();
}
}
消息消费者MessageConsumer.java
public class MessageConsumer {
private String topic;
private int partitionsNum;
private String topicConsumerGroup;
private MessageExecutor executor;
private ConsumerConnector connector;
private ExecutorService threadPool;
public MessageConsumer(String topic, int partitionsNum,String topicConsumerGroup, MessageExecutor executor){
this.topic = topic;
this.executor = executor;
this.partitionsNum = partitionsNum;
this.topicConsumerGroup = topicConsumerGroup;
createConsumerConsumer();
}
/**
* 初始化消息消费者,创建connector
* @author fuyuwei
* 2017年6月10日 下午11:02:26
* @return
*/
private boolean createConsumerConsumer() {
try{
Properties properties = new Properties();
properties.load(new ClassPathResource("consumer.properties").getInputStream());
properties.put("group.id",topicConsumerGroup);
ConsumerConfig config=new ConsumerConfig(properties);
connector=Consumer.createJavaConsumerConnector(config);
return true;
}catch (IOException e) {
BizLogger.error(e, "MessageConsumer","init kafka consumer properties error");
}
return false;
}
/**
* 接收消息,并启动线程放到线程池执行
* @author fuyuwei
* 2017年6月10日 下午11:02:51
* @throws Exception
*/
public void receiveMessage() throws Exception{
Map topics = new HashMap();
topics.put(topic, partitionsNum);
Map>> streams = connector.createMessageStreams(topics);
List> partitions = streams.get(topic);
threadPool = Executors.newFixedThreadPool(partitionsNum);
for(KafkaStream partition : partitions){
threadPool.submit(new ReceiverMessageRunner(partition));
}
}
public void close(){
try{
if (threadPool != null)
threadPool.shutdownNow();
}catch(Exception e){
BizLogger.error(e, "MessageConsumer","close fail");
}finally{
if (connector != null)
connector.shutdown();
}
}
private class ReceiverMessageRunner implements Runnable{
private KafkaStream partition;
public ReceiverMessageRunner(KafkaStream partition) {
this.partition = partition;
}
public void run(){
ConsumerIterator it = partition.iterator();
while(it.hasNext()){
MessageAndMetadata item = it.next();
executor.execute(item.message());
}
}
}
}
执行消息的保存操作MessageConsumerExecutor.java
public class MessageConsumerExecutor implements MessageExecutor {
@Override
public void execute(byte[] message ) {
try {
BizLogger.info("ReceiverMessageExecutor","start Resolve message");
String random = randomString();
int totalLength = message.length;
if(totalLength <= 4 ){
BizLogger.info("message length is not correct");
}
byte[] header = new byte[4];// 4个字节的消息头
System.arraycopy(message, 0, header, 0, 4);
int headerLength = Utility.byte2Int(header);
if(headerLength >= totalLength){
BizLogger.info("message header is not correct","headerLength",headerLength,"totalLength",totalLength);
return;
}
byte[] headerMessage = new byte[headerLength];
System.arraycopy(message, 4, headerMessage, 0, headerLength);
BizLogger.info("start parse headerMessage");
NYMobStatHeader mobheader = NYMobStatHeader.parseFrom(headerMessage);
BizLogger.info("header",mobheader.getAppVer(),mobheader.getSysVer(),mobheader.getSdkVer(),mobheader.getDeviceName(),mobheader.getTelcom(),mobheader.getImei(),mobheader.getNetworkType(),mobheader.getAppId(),mobheader.getUserId(),random + mobheader.getFileName());
int currentLength = 4 + headerLength;
while (currentLength < totalLength) {
byte[] bodyMessageFlag = new byte[4];// 4个字节代表消息体的长度
System.arraycopy(message, currentLength, bodyMessageFlag, 0, 4);
int bodyLength = Utility.byte2Int(bodyMessageFlag);
if(bodyLength >= totalLength){
BizLogger.info("message body is not correct");
return;
}
byte[] bodyMessage = new byte[bodyLength];
currentLength = currentLength + 4 ;
System.arraycopy(message, currentLength, bodyMessage, 0, bodyLength);
currentLength = currentLength + bodyLength;
NYMobStatModel statModel = NYMobStatModel.parseFrom(bodyMessage);
Map maps = statModel.getEventAttributesMap();
StringBuffer keys = new StringBuffer();
if(maps != null){
Set keySet=maps.keySet();
Iterator iterator=keySet.iterator();
while(iterator.hasNext()){
String key=iterator.next();
String value = maps.get(key);
keys.append(key).append(":").append(value).append(",");
}
}
BizLogger.info("body",statModel.getDataType(),statModel.getCtime(),statModel.getEventId(),statModel.getEventLabel(),keys.toString(),statModel.getPageId(),statModel.getFromPageId(),statModel.getUserId(),random + mobheader.getFileName());
}
} catch (InvalidProtocolBufferException e) {
BizLogger.info("protobuff parse fail ");
ErrorMessageLogger.info("ReceiverMessageExecutor","protobuff parse fail");
}catch (Exception e) {
BizLogger.info("parse fail ");
ErrorMessageLogger.info("ReceiverMessageExecutor","parse fail");
}
}
public static String randomString(){
String s = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz";
char[] c = s.toCharArray();
Random random = new Random();
StringBuffer buffer = new StringBuffer();
for(int i = 0;i< 5;i++){
buffer.append(c[random.nextInt(c.length)]);
}
return buffer.toString();
}
}
定义保存消息操作接口类MessageExecutor.java
public interface MessageExecutor {
public void execute(byte[] message) ;
}
发送消息接口类MessageService.java
public interface MessageService {
/**
* 发送消息
* @param message 消息
* @return
* @throws BusinessException
*/
public boolean sendMessage(String topic,Message message) throws Exception;
/**
* 批量发送消息
* @param messages 消息集合
* @return
* @throws BusinessException
*/
public boolean sendMessage(String topic,Collection messages)throws Exception;
}
序列化工具类GoogleprotobufUtils.java和往前台返回json信息的工具类RespUtil.java
public class RespUtil {
/**
*
* @author fuyuwei
* 2017年6月10日 下午8:23:41
* @param resp
* @param msg
* @param cachetime
* @throws IOException
*/
public static void responJson(HttpServletResponse resp, JsonMsg msg,int cachetime) throws IOException {
resp.setHeader("Access-Control-Allow-Origin", "*");
if (cachetime == 0) {
resp.setHeader("Cache-Control", "no-cache");
resp.setHeader("Pragma", "no-cache");
} else {
resp.setHeader("Cache-Control",
(new StringBuilder()).append("max-age=").append(cachetime)
.toString());
}
resp.setContentType("application/json;charset=utf-8");
resp.getWriter().write(msg.toString());
resp.getWriter().close();
}
}
消费者配置文件consumer.properties
zookeeper.connect=127.0.01:2181
# timeout in ms for connecting to zookeeper
zookeeper.session.timeout.ms=20000
zookeeper.connectiontimeout.ms=1000000
zookeeper.sync.time.ms=20000
auto.commit.enable=true
auto.commit.interval.ms=1000
queued.max.message.chunks=50
rebalance.max.retries=5
# 最大取多少块缓存到消费者(默认10)
queued.max.message.chunks=50
# 每次feth将得到多条消息,此值为总大小,提升此值,将会消耗更多的consumer端内存
fetch.min.bytes=6553600
# 当消息的尺寸不足时,server阻塞的时间,如果超时,消息将立即发送给consumer
fetch.wait.max.ms=5000
socket.receive.buffer.bytes=655360
auto.offset.reset=largest
# 指定序列化处理类(mafka client API调用说明-->3.序列化约定wiki),默认为kafka.serializer.DefaultDecoder,即byte[]
derializer.class=kafka.serializer.DefaultDecoder
生产者配置producer.properties
bootstrap.servers=127.0.01:9092
partitioner.class=org.apache.kafka.clients.producer.internals.DefaultPartitioner
value.serializer=org.apache.kafka.common.serialization.ByteArraySerializer
key.serializer=org.apache.kafka.common.serialization.StringSerializer
buffer.memory=33554432
linger.ms=0
acks=1
request.timeout.ms=10000
Spring文件配置spring-mvc.xml
json=application/json
xml=application/xml