您好,登录后才能下订单哦!
密码登录
登录注册
点击 登录注册 即表示同意《亿速云用户服务条款》
flume架构图
单节点flume配置
flume-1.4.0 启动flume
bin/flume-ng agent --conf ./conf -f conf/flume-conf.properties -Dflume.root.logger=DEBUG,console -n agent
-n表示配置文件中agent的名字
agent.sources = r1 agent.sinks = s1 agent.channels = c1 agent.sources.r1.channels = c1 agent.sinks.s1.channel = c1 #Describe/configure the sources agent.sources.r1.type = exec agent.sources.r1.command = tail -F /home/flume/loginfo #Use a channel which buffers events in memory agent.channels.c1.type = memory agent.channels.c1.capacity = 1000 #Event agent.channels.c1.transactionCapacity = 100 agent.sinks.s1.type = logger
flume-1.4.0 + kafka-0.7.2+hdfs flume配置
agent.sources = r1 agent.sinks = s_kafka s_hdfs agent.channels = c_kafka c_hdfs agent.sources.r1.channels = c_kafka c_hdfs agent.sources.r1.type = exec #下面这个脚本tail某个日志 agent.sources.r1.command = tail -F /home/flume/loginfo agent.channels.c_kafka.type = memory agent.channels.c_hdfs.type = memory agent.sinks.s_kafka.type = com.sink.FirstkafkaSink agent.sinks.s_kafka.channel = c_kafka #kafka需要连接zk,写入broker数据 agent.sinks.s_kafka.zkconnect = localhost:2181 agent.sinks.s_kafka.topic = test agent.sinks.s_kafka.serializer.class = kafka.serializer.StringEncoder agent.sinks.s_kafka.metadata.broker.list = localhost:9092 #配置文件server.properties agent.sinks.s_kafka.custom.encoding = UTF-8 agent.sinks.s_hdfs.type = hdfs agent.sinks.s_hdfs.channel = c_hdfs #默认端口8020 agent.sinks.s_hdfs.hdfs.path = hdfs://localhost:9000/root/source agent.sinks.s_hdfs.hdfs.filePrefix = events- agent.sinks.s_hdfs.hdfs.fileType = DataStream agent.sinks.s_hdfs.hdfs.writeFormat = Text agent.sinks.s_hdfs.hdfs.rollCount = 30 #达到某一数值记录生成文件 agent.sinks.s_hdfs.hdfs.rollSize = 0 agent.sinks.s_hdfs.hdfs.rollInterval = 0 agent.sinks.s_hdfs.hdfs.useLocalTimeStamp = true agent.sinks.s_hdfs.hdfs.idleTimeout = 51 agent.sinks.s_hdfs.hdfs.threadsPoolSize = 2
免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。