使用flume将数据sink到kafka

Posted 瓶子xf

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了使用flume将数据sink到kafka相关的知识,希望对你有一定的参考价值。

flume采集过程:

 

#说明:案例是flume监听目录/home/hadoop/flume_kafka采集到kafka;

启动集群

启动kafka,

启动agent,
flume-ng agent -c . -f /home/hadoop/flume-1.7.0/conf/myconf/flume-kafka.conf -n a1 -Dflume.root.logger=INFO,console
开启消费者
kafka-console-consumer.sh --zookeeper hdp-qm-01:2181 --from-beginning --topic mytopic

生产数据到kafka
数据目录:
vi /home/hadoop/flume_hbase/word.txt
12345623434

配置文件
vi flume-kafka.conf
#Name the components on this agent
a1.sources = r1
a1.sinks = k1
a1.channels = c1

#Describe/configure the source
a1.sources.r1.type = spooldir
a1.sources.r1.spoolDir=/home/hadoop/flume_kafka

# Describe the sink
a1.sinks.k1.type = org.apache.flume.sink.kafka.KafkaSink
a1.sinks.k1.kafka.topic = mytopic
a1.sinks.k1.kafka.bootstrap.servers = hdp-qm-01:9092
a1.sinks.k1.kafka.flumeBatchSize = 20
a1.sinks.k1.kafka.producer.acks = 1
a1.sinks.k1.kafka.producer.linger.ms = 1
a1.sinks.ki.kafka.producer.compression.type = snappy

# Use a channel which buffers events in memory
a1.channels.c1.type = memory
a1.channels.c1.capacity = 1000
a1.channels.c1.transactionCapacity = 100

# Bind the source and sink to the channel
a1.sources.r1.channels = c1
a1.sinks.k1.channel = c1

以上是关于使用flume将数据sink到kafka的主要内容,如果未能解决你的问题,请参考以下文章

flume 自定义 hbase sink

Flume简介与使用——Kafka Sink消费数据之Kafka安装

使用flume sink hdfs小文件优化以及HDFS小文件问题分析和解决

flume从Kafka消费数据到HDFS

flume从kafka读取数据到hdfs中的配置

Flume整合Kafka(基于kerberos认证)——完成实时数据采集