Flume配置Replicating Channel Selector
1 官网内容
上面的配置是r1获取到的内容会同时复制到c1 c2 c3 三个channel里面
2 详细配置信息
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 | # Name the components on this agent a1.sources = r1 a1.sinks = k1 k2 a1.channels = c1 c2 # Describe/configure the source a1.sources.r1.type = exec a1.sources.r1.command = tail -F /tmp/logs/cmcc.log # Describe the sink a1.sinks.k1.type = org.apache.flume.sink.kafka.KafkaSink a1.sinks.k1.topic = cmcc1 a1.sinks.k1.brokerList = hdp1: 9092 ,hdp2: 9092 ,hdp3: 9092 a1.sinks.k1.requiredAcks = 1 a1.sinks.k1.batchSize = 20 a1.sinks.k2.type = org.apache.flume.sink.kafka.KafkaSink a1.sinks.k2.topic = cmcc2 a1.sinks.k2.brokerList = hdp1: 9092 ,hdp2: 9092 ,hdp3: 9092 a1.sinks.k2.requiredAcks = 1 a1.sinks.k1.batchSize = 20 a1.sinks.k2.type = org.apache.flume.sink.kafka.KafkaSink a1.sinks.k2.topic = cmcc2 a1.sinks.k2.brokerList = hdp1: 9092 ,hdp2: 9092 ,hdp3: 9092 a1.sinks.k2.requiredAcks = 1 a1.sinks.k2.batchSize = 20 # Use a channel which buffers events in memory a1.channels.c1.type = memory a1.channels.c1.capacity = 1000 a1.channels.c1.transactionCapacity = 100 a1.channels.c2.type = file a1.channels.c2.checkpointDir = /tmp/flume/checkpoint a1.channels.c2.dataDirs = /tmp/flume/data # Bind the source and sink to the channel a1.sources.r1.channels = c1 c2 # set channel for sinks a1.sinks.k1.channel = c1 a1.sinks.k2.channel = c2 # # a1.sources.r1.selector.type = replicating # |
3 查看消费情况
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 | topic cmcc1的消费情况 kafka-console-consumer.sh --zookeeper hdp1: 2181 ,hdp2: 2181 ,hdp3: 2181 /kafka1. 1.0 --topic cmcc1 --from-beginning Using the ConsoleConsumer with old consumer is deprecated and will be removed in a future major release. Consider using the new consumer by passing [bootstrap-server] instead of [zookeeper]. hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy zhangjin xxxx yyyy topic cmcc2的消费情况 /tmp/logs]#kafka-console-consumer.sh --zookeeper hdp1: 2181 ,hdp2: 2181 ,hdp3: 2181 /kafka1. 1.0 --topic cmcc2 --from-beginning Using the ConsoleConsumer with old consumer is deprecated and will be removed in a future major release. Consider using the new consumer by passing [bootstrap-server] instead of [zookeeper]. hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy hello world java scala hadoop zhangjin xxxx yyyy zhangjin xxxx yyyy |
4 查看tail的文件内容
1 2 3 4 5 6 7 8 9 10 11 12 | hello world java scala hadoop zhangjin xxxx yyyy zhangjin xxxx yyyy |
4 总结
应该是启动了两次的原因,实际上是把文件重复两次的发送到了每个sink里面,实现了实验要求
分类:
Flume
· 理解Rust引用及其生命周期标识(上)
· 浏览器原生「磁吸」效果!Anchor Positioning 锚点定位神器解析
· 没有源码,如何修改代码逻辑?
· 一个奇形怪状的面试题:Bean中的CHM要不要加volatile?
· [.NET]调用本地 Deepseek 模型
· Blazor Hybrid适配到HarmonyOS系统
· Obsidian + DeepSeek:免费 AI 助力你的知识管理,让你的笔记飞起来!
· 解决跨域问题的这6种方案,真香!
· 一套基于 Material Design 规范实现的 Blazor 和 Razor 通用组件库
· 5. Nginx 负载均衡配置案例(附有详细截图说明++)