/** * Copyright 2015 StreamSets Inc. * * Licensed under the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.streamsets.datacollector.flume.standalone; import com.google.common.io.Resources; import com.streamsets.datacollector.base.PipelineOperationsStandaloneIT; import org.apache.flume.Channel; import org.apache.flume.ChannelSelector; import org.apache.flume.Context; import org.apache.flume.Transaction; import org.apache.flume.channel.ChannelProcessor; import org.apache.flume.channel.MemoryChannel; import org.apache.flume.channel.ReplicatingChannelSelector; import org.apache.flume.conf.Configurables; import org.apache.flume.source.AvroSource; import org.junit.AfterClass; import org.junit.BeforeClass; import java.net.URI; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; import java.util.ArrayList; import java.util.List; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; public class FlumeDestinationPipelineOperationsIT extends PipelineOperationsStandaloneIT { private static AvroSource source; private static Channel ch; private static ExecutorService executorService; @BeforeClass public static void beforeClass() throws Exception { //setup flume to write to source = new AvroSource(); ch = new MemoryChannel(); Configurables.configure(ch, new Context()); Context context = new Context(); //This should match whats present in the pipeline.json file context.put("port", String.valueOf(9050)); context.put("bind", "localhost"); Configurables.configure(source, context); List<Channel> channels = new ArrayList<>(); channels.add(ch); ChannelSelector rcs = new ReplicatingChannelSelector(); rcs.setChannels(channels); source.setChannelProcessor(new ChannelProcessor(rcs)); source.start(); PipelineOperationsStandaloneIT.beforeClass(getPipelineJson()); //read from flume memory channel every second, otherwise the channel fills up and there is no more data coming in executorService = Executors.newSingleThreadExecutor(); executorService.submit(new Runnable() { @Override public void run() { Transaction transaction; while (true) { transaction = ch.getTransaction(); transaction.begin(); ch.take(); transaction.commit(); transaction.close(); } } }); } @AfterClass public static void afterClass() throws Exception { executorService.shutdownNow(); source.stop(); ch.stop(); PipelineOperationsStandaloneIT.afterClass(); } private static String getPipelineJson() throws Exception { URI uri = Resources.getResource("flume_destination_pipeline_operations.json").toURI(); String pipelineJson = new String(Files.readAllBytes(Paths.get(uri)), StandardCharsets.UTF_8); return pipelineJson; } @Override protected String getPipelineName() { return "flume_destination_pipeline"; } @Override protected String getPipelineRev() { return "0"; } @Override protected void postPipelineStart() { } }