How to Change Hadoop Daemon log4j.properties
来源:互联网 发布:服务器上的java 编辑:程序博客网 时间:2024/06/15 06:18
refer:
https://discuss.pivotal.io/hc/en-us/articles/202296718-How-to-Change-Hadoop-Daemon-log4j-properties
Refer to the following Java Docs for log4j and sample configuration params quick reference.
DailyRollingFileAppender_DRFA## Daily Rolling File Appender# Rollover at midnightlog4j.appender.DRFA=org.apache.log4j.DailyRollingFileAppenderlog4j.appender.DRFA.File=${hadoop.log.dir}/${hadoop.log.file}log4j.appender.DRFA.DatePattern=.yyyy-MM-ddlog4j.appender.DRFA.layout=org.apache.log4j.PatternLayoutlog4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
RollingFileAppender_RFA## Rolling File Appender - cap space usage at 256mb.#hadoop.log.maxfilesize=256MBhadoop.log.maxbackupindex=20log4j.appender.RFA=org.apache.log4j.RollingFileAppenderlog4j.appender.RFA.File=${hadoop.log.dir}/${hadoop.log.file}log4j.appender.RFA.MaxFileSize=${hadoop.log.maxfilesize}log4j.appender.RFA.MaxBackupIndex=${hadoop.log.maxbackupindex}log4j.appender.RFA.layout=org.apache.log4j.PatternLayoutlog4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
FileAppender## File Appender#log4j.appender.FA=org.apache.log4j.FileAppenderlog4j.appender.FA.File=${hive.log.dir}/${hive.log.file}log4j.appender.FA.layout=org.apache.log4j.PatternLayoutlog4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
ConsoleAppender## console appender options#log4j.appender.console=org.apache.log4j.ConsoleAppenderlog4j.appender.console.target=System.errlog4j.appender.console.layout=org.apache.log4j.PatternLayoutlog4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%nlog4j.appender.console.encoding=UTF-8
Refer below to the Root logger environmental variables quick reference.
VariableServicesWhere to OverrideHADOOP_ROOT_LOGGER=INFO,RFANamenodeJournalnode
ZKFC
Datanode
Secondary Namenode/etc/gphd/hadoop/conf/hadoop-env.shHADOOP_MAPRED_ROOT_LOGGER=INFO,RFA MapReduce History Server /etc/gphd/hadoop/conf/mapred-env.shYARN_ROOT_LOGGER=INFO,RFA
Resourcemanager
Nodemanager
hbase master
HBase regionserver
Where is the log directory defined?
All services will have the $<SERVICE>_LOG_DIR variable defined in /etc/default/<service>. For example, the datanode service has "HADOOP_LOG_DIR=/var/log/gphd/hadoop-hdfs". So all logs will be found in "/var/log/gphd/hadoop-hdfs" for the datanode service.
[gpadmin@hdw1 ~]$ cat /etc/default/hadoop-hdfs-datanode | egrep ^exportexport HADOOP_PID_DIR=/var/run/gphd/hadoop-hdfsexport HADOOP_LOG_DIR=/var/log/gphd/hadoop-hdfsexport HADOOP_NAMENODE_USER=hdfsexport HADOOP_SECONDARYNAMENODE_USER=hdfsexport HADOOP_DATANODE_USER=hdfsexport HADOOP_IDENT_STRING=hdfs
Namenode, Journalnode, ZKFC, Datanode, and Secondary Namenode daemons
These daemons source their log4j settings from the following location /etc/gphd/hadoop/conf/log4j.properties
The HADOOP_ROOT_LOGGER environmental variable is used to control the default logger and is sourced in file "/usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh" which sets the root logger to RollingFileAppender by default. This can be overridden here: /etc/gphd/hadoop/conf/hadoop-env.sh
export HADOOP_ROOT_LOGGER=INFO,RFA
Audit Logging
Audit logging uses the DRFAS as per the "hadoop.security.logger" setting configured by /etc/gphd/hadoop/conf/hadoop-env.sh with HADOOP_NAMENODE_OPTS, HADOOP_DATANODE_OPTS, and HADOOP_SECONDARYNAMENODE_OPTS environmental variables.
Mapreduce History server
Mapreduce History service sources log4j settings from /etc/gphd/hadoop/conf/log4j.properties
HADOOP_MAPRED_ROOT_LOGGER environmental variable is used to control the default logger and is sourced in file "/usr/lib/gphd/hadoop-mapreduce/sbin/mr-jobhistory-daemon.sh" which sets the mapreduce history server logger to RollingFileAppender by default. This can be overridden here /etc/gphd/hadoop/conf/mapred-env.sh
export HADOOP_MAPRED_ROOT_LOGGER=INFO,RFA
Resource Manager, Nodemanager
These daemons source their log4j settings from /etc/gphd/hadoop/conf/log4j.properties
YARN_ROOT_LOGGER environmental variable is used to control the default logger and is sourced in file "/usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh" which sets the default logger to RollingFileAppender. This can be overridden in /etc/gphd/hadoop/conf/yarn-env.sh
export YARN_ROOT_LOGGER=INFO,RFA
Zookeeper
Zookeeper sources log4j settings from /etc/gphd/zookeeper/conf/log4j.properties
The ZOO_LOG4J_PROP environmental variable is used to control the default logger and is sourced in file "/usr/bin/zookeeper-server" which sets the default logger to RollingFileAppender. This can be overridden by exporting this value in the following file location /etc/gphd/zookeeper/conf/java.env
export ZOO_LOG4J_PROP=INFO,ROLLINGFILE
Hbase Master, Hbase Regionserver
These daemons source their log4j settings from /etc/gphd/hbase/conf/log4j.properties
The HBASE_ROOT_LOGGER environmental variable is used to control the default logger and is sourced in file "/usr/lib/gphd/hbase/bin/hbase-daemon.sh" which sets the default logger to RollingFileAppender. this can be overridden in /etc/gphd/hbase/conf/hbase-env.sh
export HBASE_ROOT_LOGGER=INFO,RFA
HIVE
Hive sources log4j settings from /etc/gphd/hive/conf/hive-log4j.properties In PHD all Hive daemon logs will source this file for hive.root.logger.
hive.root.logger=WARN,DRFA
hive.log.dir=/tmp/${user.name}
hive.log.file=hive.log
The following file location "/etc/init.d/hive-server" is used to start the Hive server and will set the Hive server log to file name "hive-server.log" and use the default hive.root.logger defined in hive-log4j.properties file. This log file will get truncated each time the Hive server daemon restarts.
NAME="hive-server"
LOG_FILE="/var/log/gphd/hive/${NAME}.log"
The following file location "/etc/init.d/hive-metastore" is used to start the Hive metastore and will set the Hive server log to the file name "hive-metastore" and uses the default hive.root.logger defined in hive-log4j.properties. This log file gets truncated each time the Hive metastore daemon restarts.
NAME="hive-metastore"
LOG_FILE="/var/log/gphd/hive/${NAME}.log"
Both the hive-server and hive-metastore daemon will log their data to "hive.log" as defined in hive-log4j.properties. The consolidated hive.log will get rotated as per hive.root.logger which is set to DRFA and defined in hive-log4j.properties.
Hive Query History Log
The history log file location is governed by "hive.querylog.location" from the hive-stie.xml. By default, this param is set to "/<hdfs-site.xml hadoop.tmp.dir>/${user.name}/"
- How to Change Hadoop Daemon log4j.properties
- How to change the Cloudera Manager Properties in postgres
- How to change the name of logfile dynamically using log4j
- 【GNU-Basic】How to daemon
- How to change the dbo
- How to monitor database change
- How to change fedora15+ runlevel
- How to Change Your Life
- How to change peoplesoft logo
- HOW TO CHANGE SYSADM PASSWORD?
- how to change Scheme name
- how to change home to another partition
- how to change CString to char *
- How to change em0 to eth0?
- How To Configure Hadoop
- How to change the timeout in FXCop
- How to change text of CBA buttons?
- How to change your shell prompt.
- App的启动过程(3)应用进程侧-主线程的处理
- lintcode_整数排序 II
- 让你的node、express服务器支持https
- GroupCoordinator分析
- 自定义RecycleView实现TV应用上的item焦点获取以及设置当前选中的item居中
- How to Change Hadoop Daemon log4j.properties
- 仿UC视频播放(iOS源代码)
- cstring和string区别
- Java回炉重造(二)静态代理
- K-means均值聚类算法的原理与实现
- Android线程池使用
- EF Code First 使用反射加载程序集中的Fluent API配置文件
- X86 PUSHF/PUSHFD/PUSHFQ 指令详解
- Android开发之调用摄像头拍照