前往小程序,Get更优阅读体验!
立即前往
首页
学习
活动
专区
工具
TVP
发布
社区首页 >专栏 >Flume-ng配置

Flume-ng配置

作者头像
星哥玩云
发布2022-06-29 19:31:50
2650
发布2022-06-29 19:31:50
举报
文章被收录于专栏:开源部署开源部署

1)简介

Flume是一个分布式、可靠、和高可用的海量日志聚合的系统,支持在系统中定制各类数据发送方,用于收集数据;同时,Flume提供对数据进行简单处理,并写到各种数据接受方(可定制)的能力。

设计目标: (1) 可靠性 当节点出现故障时,日志能够被传送到其他节点上而不会丢失。Flume提供了三种级别的可靠性保障,从强到弱依次分别为:end-to-end(收到数据agent首先将event写到磁盘上,当数据传送成功后,再删除;如果数据发送失败,可以重新发送。),Store on failure(当数据接收方crash时,将数据写到本地,待恢复后,继续发送),Best effort(数据发送到接收方后,不会进行确认)。 (2) 可扩展性 Flume采用了三层架构,分别为agent,collector和storage,每一层均可以水平扩展。其中,所有agent和collector由master统一管理,这使得系统容易监控和维护,且master允许有多个(使用ZooKeeper进行管理和负载均衡),这就避免了单点故障问题。 (3) 可管理性 所有agent和colletor由master统一管理,这使得系统便于维护。多master情况,Flume利用ZooKeeper和gossip,保证动态配置数据的一致性。用户可以在master上查看各个数据源或者数据流执行情况,且可以对各个数据源配置和动态加载。Flume提供了web 和shell script command两种形式对数据流进行管理。 (4) 功能可扩展性 用户可以根据需要添加自己的agent,collector或者storage。此外,Flume自带了很多组件,包括各种agent(file, syslog等),collector和storage(File,HDFS,HBase等)。

2)配置

之前配置过Hadoop和hbase,所以需要先将hadoop和hbase启动,才能将文件写入hdfs和hbase。hadoop-2.2.0和hbase-0.96.0的配置分别参考《Ubuntu和CentOS中分布式配置Hadoop-2.2.0》 http://www.linuxidc.com/Linux/2014-01/95799.htm 和《CentOS分布式环境安装HBase-0.96.0》 http://www.linuxidc.com/Linux/2014-01/95801.htm 。

本次配置环境为两台装有centos 的测试集群。主机名为master的机器负责收集日志,主机名为node的机器负责日志的写入,本次配置的写入方式有三种:写入普通目录,写入hdfs。

首先下载flume-ng的二进制压缩文件。地址:http://flume.apache.org/download.html。下载好后,解压文件。首先编辑/etc/profile文件,在其中添加如下几行:

  1. export FLUME_HOME=/home/aaron/apache-flume-1.4.0-bin
  2. export FLUME_CONF_DIR=$FLUME_HOME/conf
  3. export PATH=$PATH:$FLUME_HOME/bin
代码语言:javascript
复制
export FLUME_HOME=/home/aaron/apache-flume-1.4.0-bin
export FLUME_CONF_DIR=$FLUME_HOME/conf
export PATH=$PATH:$FLUME_HOME/bin

添加好之后记得运行$ souce /etc/profile命令使修改生效。

在master的flume文件夹的conf目录中,新建一个flume-master.conf文件,内容如下:

  1. agent.sources = seqGenSrc
  2. agent.channels = memoryChannel
  3. agent.sinks = remoteSink
  4. # For each one of the sources, the type is defined
  5. agent.sources.seqGenSrc.type = exec
  6. agent.sources.seqGenSrc.command = tail -F /home/aaron/test
  7. # The channel can be defined as follows.
  8. agent.sources.seqGenSrc.channels = memoryChannel
  9. # Each sink's type must be defined
  10. agent.sinks.loggerSink.type = logger
  11. #Specify the channel the sink should use
  12. agent.sinks.loggerSink.channel = memoryChannel
  13. # Each channel's type is defined.
  14. agent.channels.memoryChannel.type = memory
  15. # Other config values specific to each type of channel(sink or source)
  16. # can be defined as well
  17. # In this case, it specifies the capacity of the memory channel
  18. agent.channels.memoryChannel.capacity = 100
  19. agent.channels.memoryChannel.keep-alive = 100
  20. agent.sinks.remoteSink.type = avro
  21. agent.sinks.remoteSink.hostname = node
  22. agent.sinks.remoteSink.port = 23004
  23. agent.sinks.remoteSink.channel = memoryChannel
代码语言:javascript
复制
agent.sources = seqGenSrc
agent.channels = memoryChannel
agent.sinks = remoteSink

# For each one of the sources, the type is defined
agent.sources.seqGenSrc.type = exec
agent.sources.seqGenSrc.command = tail -F /home/aaron/test

# The channel can be defined as follows.
agent.sources.seqGenSrc.channels = memoryChannel

# Each sink's type must be defined
agent.sinks.loggerSink.type = logger

#Specify the channel the sink should use
agent.sinks.loggerSink.channel = memoryChannel

# Each channel's type is defined.
agent.channels.memoryChannel.type = memory

# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100
agent.channels.memoryChannel.keep-alive = 100

agent.sinks.remoteSink.type = avro
agent.sinks.remoteSink.hostname = node
agent.sinks.remoteSink.port = 23004
agent.sinks.remoteSink.channel = memoryChannel

在node机器上也将/etc/profile文件添加上面的配置。然后,在conf中新建一个flume-node.conf文件,修改如下:

  1. agent.sources = seqGenSrc1
  2. agent.channels = memoryChannel
  3. #agent.sinks = fileSink
  4. agent.sinks = <SPANstyle="FONT-FAMILY: Arial, Helvetica, sans-serif">fileSink</SPAN>
  5. # For each one of the sources, the type is defined
  6. agent.sources.seqGenSrc1.type = avro
  7. agent.sources.seqGenSrc1.bind = node
  8. agent.sources.seqGenSrc1.port = 23004
  9. # The channel can be defined as follows.
  10. agent.sources.seqGenSrc1.channels = memoryChannel
  11. # Each sink's type must be defined
  12. agent.sinks.loggerSink.type = logger
  13. #Specify the channel the sink should use
  14. agent.sinks.loggerSink.channel = memoryChannel
  15. # Each channel's type is defined.
  16. agent.channels.memoryChannel.type = memory
  17. # Other config values specific to each type of channel(sink or source)
  18. # can be defined as well
  19. # In this case, it specifies the capacity of the memory channel
  20. agent.channels.memoryChannel.capacity = 100
  21. agent.channels.memoryChannel.keep-alive = 100
  22. agent.sources.flieSink.type = avro
  23. agent.sources.fileSink.channel = memoryChannel
  24. agent.sources.fileSink.sink.directory = /home/aaron/
  25. agent.sources.fileSink.serializer.appendNewline = true
代码语言:javascript
复制
agent.sources = seqGenSrc1
agent.channels = memoryChannel
#agent.sinks = fileSink
agent.sinks = fileSink

# For each one of the sources, the type is defined
agent.sources.seqGenSrc1.type = avro 
agent.sources.seqGenSrc1.bind = node
agent.sources.seqGenSrc1.port = 23004 

# The channel can be defined as follows.
agent.sources.seqGenSrc1.channels = memoryChannel

# Each sink's type must be defined
agent.sinks.loggerSink.type = logger

#Specify the channel the sink should use
agent.sinks.loggerSink.channel = memoryChannel

# Each channel's type is defined.
agent.channels.memoryChannel.type = memory

# Other config values specific to each type of channel(sink or source)
# can be defined as well
# In this case, it specifies the capacity of the memory channel
agent.channels.memoryChannel.capacity = 100
agent.channels.memoryChannel.keep-alive = 100

agent.sources.flieSink.type = avro
agent.sources.fileSink.channel = memoryChannel
agent.sources.fileSink.sink.directory = /home/aaron/
agent.sources.fileSink.serializer.appendNewline = true

在master上面运行命令:

  1. $ bin/flume-ng agent --conf ./conf/ -f conf/flume-maste.conf -Dflume.root.logger=DEBUG,console -n agent
代码语言:javascript
复制
$ bin/flume-ng agent --conf ./conf/ -f conf/flume-maste.conf -Dflume.root.logger=DEBUG,console -n agent

在node上运行命令:

  1. $ bin/flume-ng agent --conf ./conf/ -f conf/flume-node.conf -Dflume.root.logger=DEBUG,console -n agent
代码语言:javascript
复制
$ bin/flume-ng agent --conf ./conf/ -f conf/flume-node.conf -Dflume.root.logger=DEBUG,console -n agent

启动之后,就可以发现两者之间可以相互通信,master上面的文件就能发送到node上,修改master上的test文件,在后面追加内容时,node也可以接收到。

如果想要将内容写入hadoop,可以将node中的flume-node.conf文件做如下修改:

  1. agent.sinks = k2
  2. agent.sinks.k2.type = hdfs
  3. agent.sinks.k2.channel = memoryChannel
  4. agent.sinks.k2.hdfs.path = hdfs://master:8089/hbase
  5. agent.sinks.k2.hdfs.fileType = DataStream
  6. agent.sinks.k2.hdfs.writeFormat = Text
代码语言:javascript
复制
agent.sinks = k2

agent.sinks.k2.type = hdfs
agent.sinks.k2.channel = memoryChannel
agent.sinks.k2.hdfs.path = hdfs://master:8089/hbase
agent.sinks.k2.hdfs.fileType = DataStream
agent.sinks.k2.hdfs.writeFormat = Text

其中,hdfs://master:8089/hbase为hadoop的hdfs文件路径。

本文参与 腾讯云自媒体分享计划,分享自作者个人站点/博客。
如有侵权请联系 cloudcommunity@tencent.com 删除

本文分享自 作者个人站点/博客 前往查看

如有侵权,请联系 cloudcommunity@tencent.com 删除。

本文参与 腾讯云自媒体分享计划  ,欢迎热爱写作的你一起参与!

评论
登录后参与评论
0 条评论
热度
最新
推荐阅读
相关产品与服务
TDSQL MySQL 版
TDSQL MySQL 版(TDSQL for MySQL)是腾讯打造的一款分布式数据库产品,具备强一致高可用、全球部署架构、分布式水平扩展、高性能、企业级安全等特性,同时提供智能 DBA、自动化运营、监控告警等配套设施,为客户提供完整的分布式数据库解决方案。
领券
问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档