log4j配置及意义 不在控制台输出的问题

#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements.  See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License.  You may obtain a copy of the License at
#
#    http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#

# Set everything to be logged to the console
log4j.rootLogger = DEBUG,A,D,E,W


log4j.appender.A=org.apache.log4j.ConsoleAppender
log4j.appender.A.Threshold = DEBUG
log4j.appender.A.layout=org.apache.log4j.PatternLayout
log4j.appender.A.layout.ConversionPattern=[%-d{yyyy-MM-dd HH:mm:ss}][%t]-[ %p][%c][%l] %m%n

# Set the default spark-shell log level to WARN. When running the spark-shell, the
# log level for this class is used to overwrite the root logger's log level, so that
# the user can have different defaults for the shell and regular Spark apps.
### 输出DEBUG 级别以上的日志到=/data/log/flexstream ###
log4j.appender.D = org.apache.log4j.DailyRollingFileAppender
log4j.appender.D.File = /data/log/flexstream/debug.log
log4j.appender.D.Append = true
log4j.appender.D.Threshold = DEBUG
log4j.appender.D.layout = org.apache.log4j.PatternLayout
log4j.appender.D.layout.ConversionPattern = [%-d{yyyy-MM-dd HH:mm:ss}][%t]-[ %p][%c][%l] %m%n

### 输出DEBUG 级别以上的日志到=/data/log/flexstream ###
log4j.appender.W = org.apache.log4j.DailyRollingFileAppender
log4j.appender.W.File = /data/log/flexstream/warn.log
log4j.appender.W.Append = true
log4j.appender.W.Threshold = WARN
log4j.appender.W.layout = org.apache.log4j.PatternLayout
log4j.appender.W.layout.ConversionPattern = [%-d{yyyy-MM-dd HH:mm:ss}][%t]-[ %p][%c][%l] %m%n

### 输出ERROR 级别以上的日志到=/data/log/flexstream/error.log ###
log4j.appender.E = org.apache.log4j.DailyRollingFileAppender
log4j.appender.E.File = /data/log/flexstream/error.log
log4j.appender.E.Append = true
log4j.appender.E.Threshold = ERROR
log4j.appender.E.layout = org.apache.log4j.PatternLayout
log4j.appender.E.layout.ConversionPattern = [%-d{yyyy-MM-dd HH:mm:ss}][%t]-[ %p][%c][%l] %m%n

# SPARK-9183: Settings to avoid annoying messages when looking up nonexistent UDFs in SparkSQL with Hive support
#log4j.logger.org.apache.hadoop.hive.metastore.RetryingHMSHandler=FATAL
#log4j.logger.org.apache.hadoop.hive.ql.exec.FunctionRegistry=ERROR

注意 log4j.rootLogger = DEBUG,A,D,E,W中,必须在一开始的地方写DEBUG,这样才可以输出到控制台

猜你喜欢

转载自blog.csdn.net/innersense/article/details/86644327
今日推荐