slf4j-logback 日志以json格式导入ELK

同事整理的,在此分享。logback,log4j2 等slf4j的日志实现都可以以json格式输出日志, 这里采用的是logback。当然也可以以文本行的格式输出,然后在logstash里通过grok解析,但是直接以json格式输出,在logstash处理时效率会高一点。

Logback 输出 Json格式日志文件

 为了让 logback 输出JSON 格式的日志文件,需要在pom.xml 加入如下依赖

    net.logstash.logback
    logstash-logback-encoder
    4.8
    runtime

logback日志配置示例

"errorFile"  class = "ch.qos.logback.core.rolling.RollingFileAppender" >
    class = "ch.qos.logback.classic.filter.LevelFilter" >
       ERROR
       ACCEPT
       DENY
   
    ${log.dir}/elk/error.log  
    class = "ch.qos.logback.core.rolling.TimeBasedRollingPolicy" 7 天的历史,最多占用 1G的空间 -->
       ${log.dir}/bak/error.%d{yyyy-MM-dd}.log
       7
       1GB
   
    class = "net.logstash.logback.encoder.LoggingEventCompositeJsonEncoder" >
       
      
         
            
               {
              "tags" : [ "errorlog" ],
              "project" "myproject" ,
              "timestamp" "%date{\"yyyy-MM-dd'T'HH:mm:ss,SSSZ\"}" ,
              "log_level" "%level" ,
              "thread" "%thread" ,
              "class_name" "%class" ,
              "line_number" "%line" ,
              "message" "%message" ,
              "stack_trace" "%exception{5}" ,
              "req_id" "%X{reqId}" ,
              "elapsed_time" "#asLong{%X{elapsedTime}}"
              }
            
         
      
   

 

 

Json 字段说明:

名称
说明
备注
 
 
 
 
 
tags 用于说明这条日志是属于哪一类日志            
timestamp
日志记录时间            
project
系统名称,该日志来自于哪个系统            
log_level
输出日志级别            
thread
输出产生日志的线程名。            
class_name
输出执行记录请求的调用者的全限定名
 
         
line_number
输出执行日志请求的行号
 
         
message
输出应用程序提供的信息            
stack_trace
异常栈信息            
req_id
请求ID,用于追踪请求 需要引入aop-logging          
elapsed_time
该方法执行时间,单位: 毫秒 需要引入aop-logging          
%X{key}: 表示该项来自于SLF4j MDC,需要引入 aop-logging
        com.cloud
         xspring-aop-logging
        0.7 . 1
 
针对web应用,在 web.xml 中加入 ReqIdFilter,该过滤器会在MDC 加入 reqId
     aopLogReqIdFilter
     class >com.github.nickvl.xspring.core.log.aop.ReqIdFilter class >
 
     aopLogReqIdFilter
     /*
 
 
or register in springboot like  this :
 
 
@Bean
public  FilterRegistrationBean getDemoFilter(){
     ReqIdFilter reqIdFilter= new  ReqIdFilter();
     FilterRegistrationBean registrationBean= new  FilterRegistrationBean();
     registrationBean.setFilter(reqIdFilter);
     List urlPatterns= new  ArrayList();
     urlPatterns.add( "/*" );
     registrationBean.setUrlPatterns(urlPatterns);
     registrationBean.setOrder( 100 );
     return  registrationBean;
}
 
如果需要记录该方法执行时间: elapsed_time,如果在该类或者方法上加入如下注解:
 
import  com.github.nickvl.xspring.core.log.aop.annotation.LogDebug;
import  com.github.nickvl.xspring.core.log.aop.annotation.LogInfo;
 
@LogInfo   // 当logger 设为level=INFO 会输出
@LogException (value = { @Exc (value = Exception. class , stacktrace =  false )}, warn = { @Exc ({IllegalArgumentException. class })})  //
当logger 设为level=error 会输出
 
针对dubbo 消费者的日志记录,dubbo消费者是通过 javassist 生成的动态类型,如果要监控该dubbo接口的传入参数,返回值,和调用时间 需要引入aop-logging,
以及在 eye-rpc包中的接口上给对应的类或方法 加上上面的注解。
dubbo 消费者的日志会输出如下配置:
 
  "com.alibaba.dubbo.common.bytecode"  level= "INFO"  additivity= "false" >
    "dubboApiFile" />
 

ElasticSearch 模板设置

curl -XPUT http: //localhost:9200/_template/log -d '{
   "mappings" : {
     "_default_" : {
       "_all" : {
         "enabled" false
       },
       "_meta" : {
         "version" "5.1.1"
       },
       "dynamic_templates" : [
         {
           "strings_as_keyword" : {
             "mapping" : {
               "ignore_above" 1024 ,
               "type" "keyword"
             },
             "match_mapping_type" "string"
           }
         }
       ],
       "properties" : {
         "@timestamp" : {
           "type" "date"
         },
         "beat" : {
           "properties" : {
             "hostname" : {
               "ignore_above" 1024 ,
               "type" "keyword"
             },
             "name" : {
               "ignore_above" 1024 ,
               "type" "keyword"
             },
             "version" : {
               "ignore_above" 1024 ,
               "type" "keyword"
             }
           }
         },
         "input_type" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "message" : {
           "norms" false ,
           "type" "text"
         },
         "offset" : {
           "type" "long"
         },
         "source" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "tags" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "type" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         }
       }
     }
   },
   "order" 0 ,
   "settings" : {
     "index.refresh_interval" "5s"
   },
   "template" "log-*"
}'
 
curl -XPUT http: //localhost:9200/_template/log-java -d '
 
{
   "mappings" : {
     "_default_" : {
       "properties" : {
         "log_level" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "project" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "thread" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "req_id" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "class_name" : {
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "line_number" : {
           "type" "long"
         },
         "exception_class" :{
           "ignore_above" 1024 ,
           "type" "keyword"
         },
         "elapsed_time" : {
           "type" "long"
         },
        
         "stack_trace" : {
           "type" "keyword"
         }
       }
     }
   },
   "order" 1 ,
   "settings" : {
     "index.refresh_interval" "5s"
   },
   "template" "log-java-*"
}'

logstatsh 设置

logstash-java-log
if  [fields][logType] ==  "java"  {
     json {
         source =>  "message"
         remove_field => [ "offset" ]
     }
     date {
         match => [ "timestamp" , "yyyy-MM-dd'T'HH:mm:ss,SSSZ" ]
         remove_field => [ "timestamp" ]
     }
     if  [stack_trace] {
          mutate {
             add_field => {  "exception_class"  =>  "%{stack_trace}"  }
         }
     }
     if  [exception_class] {
          mutate {
             gsub => [
                 "exception_class" "\n" "" ,
                 "exception_class" ":.*" ""
             ]
         }
     }
}

filebeat 设置

filebeat.yml
filebeat.prospectors:
- input_type: log
   paths:
     - /eyebiz/logs/eyebiz-service/elk/*.log   # eyebiz-service 日志
     - /eyebiz/logs/eyebiz-web/elk/*.log       # eyebiz-web 日志
   fields:
     logType:  "java"
     docType:  "log-java-dev"

 

你可能感兴趣的:(slf4j-logback 日志以json格式导入ELK)