2016-03-31 14 views
0

我在Ubuntu工作站上运行Flume 1.5代理,该工作站从各种设备收集日志,并将日志重新格式化为逗号分隔文件,行数很长。收集并重新格式化日志后,将它们放入spool目录,Flume Agent将日志文件发送到运行Flume代理的Hadoop服务器以接受日志文件并将它们放入HDFS目录中。Flume在连续2048个字符后添加换行

一切工作正常,除了当Flume发送文件到HDFS目录时,每行中每2048个字符后有行提供。

下面是我的flume配置文件。 是否有设置让水槽不插入换行符?

#On Ubuntu Workstation 
#list sources, sinks and channels in the agent 
agent.sources = axon_source 
agent.channels = memorychannel 
agent.sinks = AvroOut 

#define flow 
agent.sources.axon_source.channels = memorychannel 
agent.sinks.AvroOut.channel = memorychannel 
agent.channels.memorychannel.type = memory 
agent.channels.memorychannel.capacity = 100000 

#source 
agent.sources.axon_source.type = spooldir 
agent.sources.axon_source.spoolDir = /home/ubuntu/workspace/logdump 
agent.sources.axon_source.decodeErrorPolicy = ignore 

#avro out 
agent.sinks.AvroOut.type = avro 
agent.sinks.AvroOut.hostname = 172.31.12.221 
agent.sinks.AvroOut.port = 41415 
agent.sinks.AvroOut.maxIoWorkers = 2 


------------------------------------------------------------ 


#On Hadoop Server 
agent.sources = AvroIn 
agent.sources.AvroIn.type = avro 
agent.sources.AvroIn.bind = 172.31.131.1 
agent.sources.AvroIn.port = 41415 
agent.sources.AvroIn.channels = MemChan1 

agent.channels = MemChan1 
agent.channels.MemChan1.type = memory 
agent.channels.MemChan1.capacity = 100000 

agent.sinks = HDFSSink 
agent.sinks.HDFSSink.type = hdfs 
agent.sinks.HDFSSink.channel = MemChan1 
agent.sinks.HDFSSink.hdfs.path = /Logs/%Y%m/ 
agent.sinks.HDFSSink.hdfs.filePrefix = axoncapture 
agent.sinks.HDFSSink.hdfs.fileSuffix = .log 
agent.sinks.HDFSSink.hdfs.minBlockReplicas = 1 
agent.sinks.HDFSSink.hdfs.rollCount = 0 
agent.sinks.HDFSSink.hdfs.rollSize = 314572800 
agent.sinks.HDFSSink.hdfs.writeFormat = Text 
agent.sinks.HDFSSink.hdfs.fileType = DataStream 
agent.sinks.HDFSSink.hdfs.useLocalTimeStamp = True 

回答

相关问题