flume-1.6.0高可用测试&&数据入Kafka-创新互联
机器列表:
站在用户的角度思考问题,与客户深入沟通,找到陆港网站设计与陆港网站推广的解决方案,凭借多年的经验,让设计与互联网技术结合,创造个性化、用户体验好的作品,建站类型包括:网站设计制作、网站设计、企业官网、英文网站、手机端网站、网站推广、域名注册、网站空间、企业邮箱。业务覆盖陆港地区。192.168.137.115 slave0 (agent) 192.168.137.116 slave1 (agent) 192.168.137.117 slave2 (agent) 192.168.137.118 slave3 (collector) 192.168.137.119 slave4 (collector)
在每个机器上创建目录
mkdir -p /home/qun/data/flume/logs
mkdir -p /home/qun/data/flume/data
mkdir -p /home/qun/data/flume/checkpoint
下载flume最新的包:
wget tar -zxvf apache-flume-1.6.0-bin.tar.gz
在slave3,slave4配置collectors
touch $FLUME_HOME/conf/server.conf
内容如下
a1.sources = r1 a1.channels = c1 a1.sinks = k1 #set channel a1.channels.c1.type = file a1.channels.c1.checkpointDir=/home/qun/data/flume/checkpoint a1.channels.c1.dataDirs=/home/qun/data/flume/data # other node,nna to nns a1.sources.r1.type = avro a1.sources.r1.bind = slave3 a1.sources.r1.port = 52020 a1.sources.r1.interceptors = i1 a1.sources.r1.interceptors.i1.type = static a1.sources.r1.interceptors.i1.key = Collector a1.sources.r1.interceptors.i1.value = SLAVE3 a1.sources.r1.channels = c1 #set sink to kafka a1.sinks.k1.type=org.apache.flume.sink.kafka.KafkaSink a1.sinks.k1.topic=mytopic a1.sinks.k1.brokerList=kafkahost:9092 a1.sinks.k1.requiredAcks=1 a1.sinks.k1.batchSize=100 a1.sinks.k1.channel=c1
在slave0,slave1,slave2配置agent
touch $FLUME_HOME/conf/client.conf
内容如下
agent1.channels = c1 agent1.sources = r1 agent1.sinks = k1 k2 #set gruop agent1.sinkgroups = g1 #set channel agent1.channels.c1.type = file agent1.channels.c1.checkpointDir=/home/qun/data/flume/checkpoint agent1.channels.c1.dataDirs=/home/qun/data/flume/data agent1.sources.r1.channels = c1 agent1.sources.r1.type = spooldir agent1.sources.r1.spoolDir=/home/qun/data/flume/logs agent1.sources.r1.fileHeader = false agent1.sources.r1.interceptors = i1 i2 agent1.sources.r1.interceptors.i1.type = static agent1.sources.r1.interceptors.i1.key = Type agent1.sources.r1.interceptors.i1.value = LOGIN agent1.sources.r1.interceptors.i2.type = timestamp # set sink1 agent1.sinks.k1.channel = c1 agent1.sinks.k1.type = avro agent1.sinks.k1.hostname = slave3 agent1.sinks.k1.port = 52020 # set sink2 agent1.sinks.k2.channel = c1 agent1.sinks.k2.type = avro agent1.sinks.k2.hostname = slave4 agent1.sinks.k2.port = 52020 #set sink group agent1.sinkgroups.g1.sinks = k1 k2 #set failover agent1.sinkgroups.g1.processor.type = failover agent1.sinkgroups.g1.processor.priority.k1 = 10 agent1.sinkgroups.g1.processor.priority.k2 = 1 agent1.sinkgroups.g1.processor.maxpenalty = 10000
在slave3,slave4上启动collecters
flume-ng agent -n a1 -c conf -f /home/qun/apache-flume-1.6.0-bin/conf/server.conf -Dflume.root.logger=DEBUG,console
在slave0,slave1,slave2上启动agent
flume-ng agent -n agent1 -c conf -f /home/qun/apache-flume-1.6.0-bin/conf/client.conf -Dflume.root.logger=DEBUG,console
测试功能
echo "hello flume">>/home/qun/data/flume/logs/test.txt collector slave3 接收到agent的日志: 16/05/26 12:44:24 INFO file.EventQueueBackingStoreFile: Start checkpoint for /home/qun/data/flume/checkpoint/checkpoint, elements to sync = 2 16/05/26 12:44:24 INFO file.EventQueueBackingStoreFile: Updating checkpoint metadata: logWriteOrderID: 1464235734894, queueSize: 0, queueHead: 0 16/05/26 12:44:24 INFO file.Log: Updated checkpoint for file: /home/qun/data/flume/data/log-3 position: 786 logWriteOrderID: 1464235734894 16/05/26 12:44:24 INFO file.Log: Removing old file: /home/qun/data/flume/data/log-1 16/05/26 12:44:24 INFO file.Log: Removing old file: /home/qun/data/flume/data/log-1.meta 16/05/26 12:44:54 INFO file.EventQueueBackingStoreFile: Start checkpoint for /home/qun/data/flume/checkpoint/checkpoint, elements to sync = 2 16/05/26 12:44:54 INFO file.EventQueueBackingStoreFile: Updating checkpoint metadata: logWriteOrderID: 1464235734901, queueSize: 0, queueHead: 0 16/05/26 12:44:54 INFO file.Log: Updated checkpoint for file: /home/qun/data/flume/data/log-3 position: 1179 logWriteOrderID: 1464235734901
测试collecters Failover
杀死slave3的flume进程,kill -9 pid
echo "hello flume">>/home/qun/data/flume/logs/test.txt collector slave4 接收到agent的日志: 16/05/26 12:08:27 INFO file.EventQueueBackingStoreFile: Start checkpoint for /home/qun/data/flume/checkpoint/checkpoint, elements to sync = 2 16/05/26 12:08:27 INFO file.EventQueueBackingStoreFile: Updating checkpoint metadata: logWriteOrderID: 1464234987484, queueSize: 0, queueHead: 0 16/05/26 12:08:27 INFO file.Log: Updated checkpoint for file: /home/qun/data/flume/data/log-3 position: 393 logWriteOrderID: 1464234987484 16/05/26 12:08:27 INFO file.LogFile: Closing RandomReader /home/qun/data/flume/data/log-1 16/05/26 12:54:38 INFO client.ClientUtils$: Fetching metadata from broker id:0,host:xiaobin,port:9092 with correlation id 4 for 1 topic(s) Set(mytopic) 16/05/26 12:54:38 INFO producer.SyncProducer: Connected to xiaobin:9092 for producing 16/05/26 12:54:38 INFO producer.SyncProducer: Disconnecting from xiaobin:9092 16/05/26 12:54:38 INFO producer.SyncProducer: Disconnecting from xiaobin:9092 16/05/26 12:54:38 INFO producer.SyncProducer: Connected to xiaobin:9092 for producing 16/05/26 12:54:57 INFO file.EventQueueBackingStoreFile: Start checkpoint for /home/qun/data/flume/checkpoint/checkpoint, elements to sync = 2 16/05/26 12:54:57 INFO file.EventQueueBackingStoreFile: Updating checkpoint metadata: logWriteOrderID: 1464234987491, queueSize: 0, queueHead: 0 16/05/26 12:54:57 INFO file.Log: Updated checkpoint for file: /home/qun/data/flume/data/log-3 position: 786 logWriteOrderID: 1464234987491 16/05/26 12:54:57 INFO file.Log: Removing old file: /home/qun/data/flume/data/log-1 16/05/26 12:54:57 INFO file.Log: Removing old file: /home/qun/data/flume/data/log-1.meta
一会儿再写····
另外有需要云服务器可以了解下创新互联scvps.cn,海内外云服务器15元起步,三天无理由+7*72小时售后在线,公司持有idc许可证,提供“云服务器、裸金属服务器、高防服务器、香港服务器、美国服务器、虚拟主机、免备案服务器”等云主机租用服务以及企业上云的综合解决方案,具有“安全稳定、简单易用、服务可用性高、性价比高”等特点与优势,专为企业上云打造定制,能够满足用户丰富、多元化的应用场景需求。
网页题目:flume-1.6.0高可用测试&&数据入Kafka-创新互联
文章源于:http://hbruida.cn/article/gjoed.html