flume+hdfs

1、安装flume以及hadoop

2、配置flume配置文件,hdfs.conf

#agent.sources.s1.type = spooldir
#agent.sources.s1.spoolDir = D:\\logs
#agent.sources.s1.fileHeader = false
#agent.sources.s1.channels=c1

agent.sources.s1.type=exec
agent.sources.s1.command=tail -F /tmp/cdz/log/test.log
agent.sources.s1.channels=c1


agent.channels.c1.type=memory
agent.channels.c1.capacity=10000
agent.channels.c1.transactionCapacity=100

agent.sinks.k1.type = hdfs
agent.sinks.k1.channel = c1
agent.sinks.k1.hdfs.path = /flume/events/%y-%m-%d/%H%M/%S
agent.sinks.k1.hdfs.filePrefix = events-
agent.sinks.k1.hdfs.round = true
agent.sinks.k1.hdfs.useLocalTimeStamp = true
agent.sinks.k1.hdfs.roundValue = 1
agent.sinks.k1.hdfs.roundUnit = minute
agent.sinks.k1.hdfs.fileType=DataStream


agent.channels = c1
agent.sources = s1
agent.sinks = k1


3、运行

sh bin/flume-ng agent --conf ./conf --conf-file ./conf/hdfs.conf --name agent -Dflume.root.logger=INFO,console


4、往/tmp/cdz/log/test.log写入数据,在hdfs的管理界面可以看到events-.*的文件,里面便是前面写入log文件的内容。

发布了65 篇原创文章 · 获赞 30 · 访问量 18万+

猜你喜欢

转载自blog.csdn.net/i792439187/article/details/53410937