logstash安装配置入kafka(配置hadoop审计日志)

来源:互联网 发布:去除单元格内重复数据 编辑:程序博客网 时间:2024/05/16 02:35

es部署情况

   - 10.183.93.129    - 10.183.93.131   - 10.183.93.132

logstash安装

#!/bin/bashcd /letvrsync -avzP 10.180.92.199::wVioz35SWO9zywesmagfOrP9XjigoF8j/james/logstash.tar.gz .tar -xzf logstash.tar.gzln -s  /letv/logstash-2.4.0 /usr/local/logstashexport LOGSTASH_HOME=/usr/local/logstashecho "export LOGSTASH_HOME=/usr/local/logstash export PATH=${LOGSTASH_HOME}/bin:$PATH" >> /root/.bashrcsource /etc/profile

nginx 配了一个json的format日志

    log_format json '{ "@timestamp": "$time_iso8601", '                         '"@fields": { '                         '"remote_addr": "$remote_addr", '                         '"remote_user": "$remote_user", '                         '"upstream_response_time": "$upstream_response_time", '                         '"request_time": "$request_time", '                         '"status": "$status", '                         '"upstream_addr": "$upstream_addr", '                         '"server_protocol": "$server_protocol", '                         '"host": "$host", '                         '"request_uri": "$request_uri", '                         '"request": "$request", '                         '"request_method": "$request_method", '                         '"http_referrer": "$http_referer", '                         '"body_bytes_sent":"$body_bytes_sent", '                         '"request_length":"$request_length", '                         '"bytes_sent":"$bytes_sent", '                         '"content_type":"$content_type", '                         '"request_body":"$request_body",'                         '"remote_port":"$remote_port",'                         '"request_body_file":"$request_body_file",'                         '"cookie_COKIE":"$cookie_COKIE",'                         '"http_x_forwarded_for": "$http_x_forwarded_for", '                         '"http_user_agent": "$http_user_agent" } }';

logstash配置文件
/etc/logstash/conf.d/lbgate.conf

input {    file {        path => "/var/log/nginx/matrix*.json.log"        codec => json        start_position => "beginning"        type => "nginx-log"    }}output {    if [type] == "nginx-log" {        elasticsearch {            hosts => ["10.183.93.129:9200"]            index => "nginx-log-%{+YYYY.MM.dd}"        }    }

写了一个入kafka的,后面再通过python-kafka消费

input {    file {        path => "/var/log/nginx/matrix*json.log"        codec => json        start_position => "beginning"        type => "nginx-log"    }}output {    if [type] == "nginx-log" {        elasticsearch {            hosts => ["10.183.93.129:9200"]            index => "nginx-log-%{+YYYY.MM.dd}"        }    }    if [type] == "nginx-log" {          kafka {              codec => json              bootstrap_servers => "bops-10-183-93-131:9092,bops-10-183-93-132:9092,bops-10-183-93-129:9092"              topic_id => "yanbo"              timeout_ms => 10000              retries => 3              client_id => "yanbo_client"          }          # stdout { codec => rubydebug }      }}

hadoop审计日志

input {      file {       type => "hdfs-audit"           path => "/data/hadoop/data12/hadoop-logs/hdfs-audit.log"           start_position => beginning           sincedb_path => "/data/hadoop/data12/hadoop-logs/logstash"       }  }  filter{      if [type] == "hdfs-audit" {         grok {             match => ["message", "ugi=(?<user>([\w\d\-]+))@|ugi=(?<user>([\w\d\-]+))/[\w\d\-.]+@|ugi=(?<user>([\w\d.\-_]+))[\s(]+"]         }      }  }  output {      if [type] == "hdfs-audit" {          kafka {              codec => plain {                  format => "%{message}"              }              bootstrap_servers => "rm1:9092,rm2:9092,test-nn1:9092,test-nn2:9092,10-140-60-50:9092"              topic_id => "hdfslog"              timeout_ms => 10000              retries => 3              client_id => "hdfs-audit"          }          # stdout { codec => rubydebug }      }  }
原创粉丝点击