Flume是一个高可用、高可靠、分布式的海量数据采集、聚合、传输的系统。
Flume作为一个日志收集工具,非常轻量级,基于一个个Flume Agent,能够构建一个复杂而强大的日志手机系统。运行在服务器上的进程将需要关注的日志文件采集到指定的存储服务,这个进程就叫Flume Agent。
本地环境
Hadoop版本:2.7.6
操作系统 : manjaro-64bit
内存 : 8GB
CPU : i7-4770S
前期准备
Flume安装包:下载链接 。
Flume安装和配置
解压Flume:
1
tar xzvf apache-flume-1.8.0-bin.tar.gz ~/apps
配置环境变量,打开~/.bashrc(使用zsh的话打开~/.zshrc),在文件末尾添加如下内容(FLUME_HOME的具体路径根据你自己解压的flume文件所在路径进行配置):
1
2
3# FLUME
export FLUME_HOME=/home/jony/apps/apache-flume-1.8.0-bin
export PATH=$FLUME_HOME/bin:$PATH使环境变量生效
1
2
3source ~/.bashrc
or
source ~/.zshrc修改Flume配置文件,将apache-flume-1.8.0-bin/conf/flume-env.sh.template重命名为flume-env.sh,打开flume-env.sh配置如下内容:
1
2
3
4# 这里如果你在安装Hadoop那里已经配置过JAVA_HOME,这里就不需要配置JAVA_HOME啦
# export JAVA_HOME=/usr/lib/jvm/java-8-oracle
export HADOOP_HOME=/home/hadoop/hadoop-current
FLUME_CLASSPATH="$HADOOP_HOME/bin:$PATH"修改apache-flume-1.8.0-bin/conf/flume-conf.properties.template为flume-conf.properties,测试Flume配置是否正确,测试命令如下:
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf.properties --name agent
打开log文件查看运行结果:
1
2
3
4
526 十二月 2018 14:08:56,172 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.sink.LoggerSink.process:95) - Event: { headers:{} body: 37 34 31 36 38 74168 }
26 十二月 2018 14:08:56,172 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.sink.LoggerSink.process:95) - Event: { headers:{} body: 37 34 31 36 39 74169 }
26 十二月 2018 14:08:56,172 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.sink.LoggerSink.process:95) - Event: { headers:{} body: 37 34 31 37 30 74170 }
26 十二月 2018 14:08:56,172 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.sink.LoggerSink.process:95) - Event: { headers:{} body: 37 34 31 37 31 74171 }
26 十二月 2018 14:08:56,172 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.sink.LoggerSink.process:95) - Event: { headers:{} body: 37 34 31 37 32 74172 }
Flume原理
Flume Agent
Flume是一个基于Flume Agent组建而成的复杂而强大的日志收集系统,而Flume Agent就是将运行在服务器上的进程的相关日志信息采集到指定存储的服务进程。
Flume Agent的组成
Flume Agent由三部分组成:Source、Channel、Sink。
source:从数据接收器接收数据,并将数据一Flume的event格式传递给一个或多个Channel,Flume提供多种数据接收方式,比如taildir,Kafka等。
Channel:Channel是一种短暂的存储容器,它将从source接收到的event格式的数据缓存起来,知道它们被sink消费掉。它在source和sink间起着一种桥梁的作用。支持的类类型有:FileChannel,Memory Channel等。
Sink:sink将数据保存到存储器中,它从channel消费数据(event)并将其传递给目标地址,目标地可能是Hive,HDFS,Hbase等。
Flume Agent基本工作流程
Flume的核心是把数据从数据源(Source)收集过来,再将收集到的数据发送到指定的目的地(Sink),为保证发送过程的正确性,在发生那个到目的地之前,会先缓存数据,用于缓存数据的通道我们就称之为Channel,待数据真正到达目的地Sink后,Flume再删除缓存的数据。
Flume Event
Flume Event是将传输数据进行封装的数据格式,是Flume进行数据传输的基本单位,如果是文本文件,通常是一行记录。
Event从source流向channel,在到sink,其本身为字节数组,可携带header信息,比如:
1 | FlumeEvent数据结构: |
Event是事务的基本单位,代表一个数据的最小完成单元。
Flume的特性
模块化设计: source、channel、sink三者可以根据业务自由组合,构建相对复杂的日志管道流。
可接入性: 支持多种主流系统和框架,如HDFS、HBase、Kafka、ES、Thrift、Avro等,都能很好地与Flume集成。
容错性: Flume中的Event是具有事务性质的,这保证了数据在源端和目的端的一致性。
可扩展性: 可以根据自己业务需要来定制实现某些组建,比如定制话的Source、Channle、Sink。
Flume使用示例
1. 监听端口,输出到控制台
在flume的conf目录下编写配置文件flume-conf-netsrc.properties,文件配置内容如下:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25agent.sources = netSrc
agent.channels = memoryChannel
agent.sinks = loggerSink
# For each one of the sources, the type is defined
agent.sources.netSrc.type = netcat
agent.sources.netSrc.bind = 0.0.0.0
agent.sources.netSrc.port = 12345
# The channel can be defined as follows.
agent.sources.netSrc.channels = memoryChannel
# Each sink's type must be defined
agent.sinks.loggerSink.type = logger
#Specify the channel the sink should use
agent.sinks.loggerSink.channel = memoryChannel
# Each channel's type is defined.
agent.channels.memoryChannel.type = memory
# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100sources类型配置为netcat,用于监听网络端口,这里bind为监听的ip地址,0.0.0.0表示监听本机。port为监听的端口。sink的类型设置为logger表示直接将监听到的信息打印到终端输出。channel选用memor类型,表示使用内存作为Channel。
启动flume:
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf-netsrc.properties --name agent -Dflume.root.logger=INFO,console
启动命令最后添加的-Dflume.root.logger=INFO,console是为了将日志输出到控制台,所以直接修改logger,使用控制台作为输出,并且日志输出级别设置为INFO。
开启telnet,向flume发送数据:
1
telnet localhost 12345
输入一些内容:
1
2
3
4
5
6
7
8hello
OK
workd
OK
workd
OK
world
OK在flume启动的终端下面查看输出:
1
2
3
42018-12-17 15:12:02,438 (SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO - org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: { headers:{} body: 68 65 6C 6C 6F 0D hello. }
2018-12-17 15:12:03,616 (SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO - org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: { headers:{} body: 77 6F 72 6B 64 0D workd. }
2018-12-17 15:12:05,047 (SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO - org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: { headers:{} body: 77 6F 72 6B 64 0D workd. }
2018-12-17 15:12:06,454 (SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO - org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: { headers:{} body: 77 6F 72 6C 64 0D world. }
2. 监听端口,输出到服务器文件
在flume的conf目录下编写配置文件flume-conf-netsrc2localfile.properties,文件配置内容如下:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27agent.sources = netSrc
agent.channels = memoryChannel
agent.sinks = fileSink
# For each one of the sources, the type is defined
agent.sources.netSrc.type = netcat
agent.sources.netSrc.bind = 0.0.0.0
agent.sources.netSrc.port = 12345
# The channel can be defined as follows.
agent.sources.netSrc.channels = memoryChannel
# Each sink's type must be defined
agent.sinks.fileSink.type = file_roll
agent.sinks.fileSink.sink.directory=/home/jony/tmp/flume_workspace/file_roll/data
agent.sinks.fileSink.sink.rollInterval = 60
#Specify the channel the sink should use
agent.sinks.fileSink.channel = memoryChannel
# Each channel's type is defined.
agent.channels.memoryChannel.type = memory
# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100这里sink的类型使用的是file_roll,表示要将最终结果输出到本地文件中,directory配置的是输出的文件路径,需要注意的是这个目录需要被提前创建,否则执行会出错。
启动flume
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf-netsrc2localfile.properties --name agent -Dflume.root.logger=INFO,console
使用telnet往对应端口发送数据:
1
telnet localhost 12345
发送数据:
1
2
3
4
5
6hello
OK
world
OK
flume
OK在指定目录下(这里是/home/jony/tmp/flume_workspace/file_roll/data)查看相关文件:
1
2
3
4$ cat 1545031636278-1
hello
world
flume60秒后,该目录下又生成了一个文件1545031636278-2.
3. 监听服务器文件,输出到HDFS(memoryChannel)
修改flume的conf目录下的flume-conf-taildir2hdfs.properties,文件配置内容如下:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32agent.sources = fileSrc
agent.channels = memoryChannel
agent.sinks = hdfsSink
# For each one of the sources, the type is defined
agent.sources.fileSrc.type = taildir
agent.sources.fileSrc.positionFile = /home/jony/tmp/flume_workspace/mem_ch/positionFile
agent.sources.fileSrc.filegroups = f1
agent.sources.fileSrc.filegroups.f1 = /home/jony/tmp/log/.*.log
# The channel can be defined as follows.
agent.sources.fileSrc.channels = memoryChannel
# Each sink's type must be defined
agent.sinks.hdfsSink.type = hdfs
agent.sinks.hdfsSink.hdfs.path = /user/jony/flume_data/mem_ch/%Y-%m-%d
agent.sinks.hdfsSink.hdfs.useLocalTimeStamp = true
agent.sinks.hdfsSink.hdfs.fileType = DataStream
agent.sinks.hdfsSink.hdfs.rollSize = 0
agent.sinks.hdfsSink.hdfs.rollCount = 0
agent.sinks.hdfsSink.hdfs.rollInterval = 60
#Specify the channel the sink should use
agent.sinks.hdfsSink.channel = memoryChannel
# Each channel's type is defined.
agent.channels.memoryChannel.type = memory
# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100修改conf/flume-env.sh,添加Hadoop的路径,Hadoop的配置可以参考我的另一篇文章:配置本地Hadoop环境 。如果你已经配置了Hadoop,这里是不用在flume-env.sh中配置任何东西的:
1
2export HADOOP_HOME=/home/jony/apps/hadoop-2.7.6
FLUME_CLASSPATH="$HADOOP_HOME/bin:$PATH"启动flume:
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf-taildir2hdfs.properties --name agent
查看HDFS上是否已经写入数据:
1
2
3$ hadoop fs -ls /user/jony/flume_data/mem_ch
Found 1 items
drwxr-xr-x - jony supergroup 0 2018-12-26 15:53 /user/jony/flume_data/mem_ch/2018-12-26
4. 监听服务器文件,输出到HDFS(fileChannel)
修改flume的conf目录下的flume-conf-taildir2hdfs.properties,文件配置内容如下:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30agent.sources = fileSrc
agent.channels = fileChannel
agent.sinks = hdfsSink
# For each one of the sources, the type is defined
agent.sources.fileSrc.type = taildir
agent.sources.fileSrc.positionFile = /home/jony/tmp/flume_workspace/hdfs_sink/file_ch/positionFile
agent.sources.fileSrc.filegroups = f1
agent.sources.fileSrc.filegroups.f1 = /home/jony/tmp/log/.*.log
# The channel can be defined as follows.
agent.sources.fileSrc.channels = fileChannel
# Each sink's type must be defined
agent.sinks.hdfsSink.type = hdfs
agent.sinks.hdfsSink.hdfs.path = /user/jony/flume_data/file_ch/%Y-%m-%d
agent.sinks.hdfsSink.hdfs.useLocalTimeStamp = true
agent.sinks.hdfsSink.hdfs.fileType = DataStream
agent.sinks.hdfsSink.hdfs.rollSize = 0
agent.sinks.hdfsSink.hdfs.rollCount = 0
agent.sinks.hdfsSink.hdfs.rollInterval = 60
#Specify the channel the sink should use
agent.sinks.hdfsSink.channel = fileChannel
# Each channel's type is defined.
agent.channels.fileChannel.type = file
agent.channels.fileChannel.checkpointDir = /home/jony/tmp/flume_workspace/hdfs_sink/file_ch/checkpoint
agent.channels.fileChannel.checkpointInterval = 60000
agent.channels.fileChannel.dataDirs = /home/jony/tmp/flume_workspace/hdfs_sink/file_ch/data启动flume:
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf-taildir2hdfs-filech.properties --name agent
查看本地checkpoint和positionFile文件:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18hdfs_sink
└── file_ch
├── checkpoint
│ ├── checkpoint
│ ├── checkpoint.meta
│ ├── inflightputs
│ ├── inflighttakes
│ └── queueset
├── data
│ ├── log-2
│ ├── log-2.meta
│ ├── log-3
│ ├── log-3.meta
│ ├── log-4
│ ├── log-4.meta.tmp
│ ├── log-5
│ └── log-5.meta
└── positionFile查看HDFS上的日志文件
1
hadoop fs -ls /user/jony/flume_data/file_ch/2018-12-26
Flume高级配置
多个FLume串联
比如下面这种情况:
Flume串联需要满足前一个agent的sink和后一个agent的source以avro的方式连接起来即可。
大型客户端生成日志,存储系统处理能力不足的情况下,可以将flume配置成层级结构:
可以在每台服务器上配置一个flume agent,然后将这些agent再串联一个公用的agent进行日志收集,这个共用的agent将收集到的日志写到hdfs上。
这里会存在一个隐患,如果公用的agent挂了,那么日志系统整个就崩溃了.
针对这种情况,可以使用flume的负载均衡模式,配置如下:
1 | #load-balance |
具体场景
需求: 采集客户端的不同服务生成的日志,根据类型发送到不同的存储服务。
针对这种情况,明显是要一个source对应多个sink,可以采用flume的扇出模式,扇出有两种方式:复制和多路复用,支持同一份source发送到不同的sink。
复制:
1
2
3
4
5
6
7
8agent_foo.sources = r1
agent_foo.channels = c1 c2 c3
agent_foo.sinks = k1 k2 k3
agent_foo.sources.r1.channels = c1 c2 c3
agent_foo.sinks.k1.channel = c1
agent_foo.sinks.k2.channel = c2
agent_foo.sinks.k3.channel = c3多路复用:
1
2
3
4
5
6
7
8
9
10
11
12
13
14agent_foo.sources = r1
agent_foo.channels = c1 c2 c3
agent_foo.sinks = k1 k2 k3
agent_foo.sources.r1.channels = c1 c2 c3
agent_foo.sinks.k1.channel = c1
agent_foo.sinks.k2.channel = c2
agent_foo.sinks.k3.channel = c3
agent_foo.sources.r1.selector.type = multiplexing
agent_foo.sources.r1.selector.header = state
agent_foo.sources.r1.selector.mapping.CZ = c1
agent_foo.sources.r1.selector.mapping.US = c2
agent_foo.sources.r1.selector.default = c3假如event内的数据格式如下:
1
2
3{"headers" : {"state" : "CZ"}, "body" : "msg1"}
{"headers" : {"state" : "US"}, "body" : "msg2"}
{"headers" : {"state" : "XX"}, "body" : "msg3"}那么msg1将分配到c1,msg2将分配到c2,msg3将分配到c3。
headers怎么设置呢?flume有内置方法:拦截器。
拦截器
拦截器的存在使用户可以向event中添加自己想要的内容。拦截器分一下几种:
Timestamp Interceptor:在event中插入当前时间戳
Host Interceptor:插入主机名
Static Interceptor:插入固定内容
Regex Filetering Interceptor:插入正则表达式匹配的内容
举个栗子
需求: 根据日志内的时间进行分类。
分析: 要根据日志内容进行分类,可以采用拦截器,在根据每条日志内的时间,添加event的header,这样在sink输出时就能根据header的不同进行不同的存储处理。
实现:
拦截器配置flume-conf-taildir2hdfs-interceptor.properties:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37agent.sources = fileSrc
agent.channels = memoryChannel
agent.sinks = hdfsSink
# For each one of the sources, the type is defined
agent.sources.fileSrc.type = taildir
agent.sources.fileSrc.positionFile = /home/jony/tmp/flume_workspace/hdfs_sink/interceptor/positionFile
agent.sources.fileSrc.filegroups = f1
agent.sources.fileSrc.filegroups.f1 = /home/jony/tmp/log/.*.log
agent.sources.fileSrc.interceptors = i1
agent.sources.fileSrc.interceptors.i1.type = regex_extractor
agent.sources.fileSrc.interceptors.i1.regex = (\\d\\d\\d\\d-\\d\\d-\\d\\d)
agent.sources.fileSrc.interceptors.i1.serializers = s1
agent.sources.fileSrc.interceptors.i1.serializers.s1.name = date
# The channel can be defined as follows.
agent.sources.fileSrc.channels = memoryChannel
# Each sink's type must be defined
agent.sinks.hdfsSink.type = hdfs
agent.sinks.hdfsSink.hdfs.path = /user/jony/flume_data/interceptor/%{date}
agent.sinks.hdfsSink.hdfs.fileType = DataStream
agent.sinks.hdfsSink.hdfs.rollSize = 0
agent.sinks.hdfsSink.hdfs.rollCount = 0
agent.sinks.hdfsSink.hdfs.rollInterval = 60
#Specify the channel the sink should use
agent.sinks.hdfsSink.channel = memoryChannel
# Each channel's type is defined.
agent.channels.memoryChannel.type = memory
# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100启动Flume;
1
./bin/flume-ng agent --conf conf --conf-file conf/flume-conf-taildir2hdfs-interceptor.properties --name agent
查看hdfs上文件:
1
2
3
4
5
6
7jony@jony-manjaro ~/apps/hadoop-2.7.6 hadoop fs -ls /user/jony/flume_data/interceptor
Found 5 items
drwxr-xr-x - jony supergroup 0 2018-12-26 19:13 /user/jony/flume_data/interceptor/2018-05-29
drwxr-xr-x - jony supergroup 0 2018-12-26 19:14 /user/jony/flume_data/interceptor/2018-05-30
drwxr-xr-x - jony supergroup 0 2018-12-26 19:14 /user/jony/flume_data/interceptor/2018-05-31
drwxr-xr-x - jony supergroup 0 2018-12-26 19:14 /user/jony/flume_data/interceptor/2018-06-01
drwxr-xr-x - jony supergroup 0 2018-12-26 19:14 /user/jony/flume_data/interceptor/2018-06-02查看本地的positionFile:
1
2/home/jony/tmp/flume_workspace/hdfs_sink/interceptor
└── positionFile
大功告成!
总结
在之前配置的Hadoop的基础上,我们安装了Flume日志收集工具,并通过四个例子,简单了解了flume的基本用法,其实在实际应用中flume的配置是比较复杂的,一般都会多级连接,并及进行负载均衡配置,将多个采集节点上的数据安全可靠地收集起来,汇聚到一个节点,然后对数据进行处理!