Flume_使用
Posted
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Flume_使用相关的知识,希望对你有一定的参考价值。
案例一: source:hive.log channel: memory sink: logger输出
拷贝一份flume-conf.properties.template改名为hive-mem-log.properties hive-mem-log.properties a1.sources = s1 a1.channels = c1 a1.sinks = k1 # defined the source a1.sources.s1.type = exec a1.sources.s1.command = tail -F /opt/cdh-5.6.3/hive-0.13.1-cdh5.3.6/logs/hive.log a1.sources.s1.shell = /bin/sh -c # defined the channel a1.channels.c1.type = memory # defined the sink a1.sinks.k1.type = logger # The channel can be defined as follows. a1.sources.s1.channels = c1 a1.sinks.k1.channel = c1 flmue目录下执行 bin/flume-ng agent -c conf/ -n al -f conf/hive-mem-log.properties -Dflume.root.logger=INFO,console 在hive端打印几条命令查看 注意flume的启动顺序和关闭顺序是不同的
案例二:source:hive.log channel: file sink: logger输出
拷贝一份flume-conf.properties.template改名为hive-file-log.properties hive-file-log.properties a1.sources = s1 a1.channels = c1 a1.sinks = k1 # defined the source a1.sources.s1.type = exec a1.sources.s1.command = tail -F /opt/cdh-5.6.3/hive-0.13.1-cdh5.3.6/logs/hive.log a1.sources.s1.shell = /bin/sh -c # defined the channel a1.channels.c1.type = file a1.channels.c1.checkpointDir = /opt/cdh-5.6.3/apache-flume-1.5.0-cdh5.3.6-bin/datas/checkp a1.channels.c1.dataDirs = /opt/cdh-5.6.3/apache-flume-1.5.0-cdh5.3.6-bin/datas/data # defined the sink a1.sinks.k1.type = logger # The channel can be defined as follows. a1.sources.s1.channels = c1 a1.sinks.k1.channel = c1 flmue目录下执行 bin/flume-ng agent -c conf/ -n a1 -f conf/hive-file-log.properties -Dflume.root.logger=INFO,console 查看自定义文件夹下数据文件
案例三:source:hive.log channel: mem sink: hdfs
拷贝一份flume-conf.properties.template改名为hive-mem-hdfs.properties hive-mem-hdfs.properties a1.sources = s1 a1.channels = c1 a1.sinks = k1 # defined the source a1.sources.s1.type = exec a1.sources.s1.command = tail -F /opt/cdh-5.6.3/hive-0.13.1-cdh5.3.6/logs/hive.log a1.sources.s1.shell = /bin/sh -c # defined the channel a1.channels.c1.type = memory a1.channels.c1.capacity = 1000 a1.channels.c1.transactionCapacity = 1000 # defined the sink a1.sinks.k1.type = hdfs a1.sinks.k1.hdfs.path = /flume/hdfs/ a1.sinks.k1.hdfs.fileType = DataStream # The channel can be defined as follows. a1.sources.s1.channels = c1 a1.sinks.k1.channel = c1 flmue目录下执行 bin/flume-ng agent -c conf/ -n a1 -f conf/hive-mem-hdfs.properties -Dflume.root.logger=INFO,console 查看HDFS下数据文件 可以发现,定义的目录不存在时,会自动被创建
以上是关于Flume_使用的主要内容,如果未能解决你的问题,请参考以下文章
任务8:08_尚硅谷_Flume案例_监控本地变化文件(需求分析)
大数据技术之_18_大数据离线平台_02_Nginx+Mysql+数据收集+Web 工程 JS/JAVA SDK 讲解+Flume 故障后-如何手动上传 Nginx 日志文件至 HDFS 上(示例代码