1、启动kafka,使用默认配置 apache
①启动zookeeper,bin/-server-start.sh -daemon config/server.properties ide
②启动kafka:bin/kafka-server-start.sh -daemon config/server.properties
③建立主题:bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic spider
④启动消费者:bin/kafka-console-consumer.sh --zookeeper localhost:2181 --topic spider --from-beginning 测试
2、配置flume并启动 ui
flume的详细配置能够看我前面的关于flume的详细配置 spa
将flume(1.6.0,自定义sink会在后面博客补充)的sink改成: server
agent1.sinks.log-sink1.channel = ch1
agent1.sinks.log-sink1.type = org.apache.flume.sink.kafka.KafkaSink
agent1.sinks.log-sink1.topic = spider
agent1.sinks.log-sink1.brokerList = 192.168.57.133:9092
agent1.sinks.log-sink1.requiredAcks = 1
agent1.sinks.log-sink1.batchSize = 1
启动flume kafka
3、测试 博客
这里的测试类仍是使用上一次log4j整合flume的例子,运行测试类,若是看到下面输出,则表示整合成功 it