/**
* Copyright 2015 StreamSets Inc.
*
* Licensed under the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.streamsets.datacollector.kafka.standalone;
import com.google.common.io.Resources;
import com.streamsets.datacollector.base.PipelineRunStandaloneIT;
import com.streamsets.pipeline.kafka.common.KafkaTestUtil;
import kafka.consumer.ConsumerIterator;
import kafka.consumer.KafkaStream;
import org.junit.After;
import org.junit.Before;
import java.net.URI;
import java.nio.charset.StandardCharsets;
import java.nio.file.Files;
import java.nio.file.Paths;
import java.util.List;
public class KafkaDestinationMultiPartitionPipelineRunIT extends PipelineRunStandaloneIT {
private static final String TOPIC = "TestKafkaDestinationMultiPartition";
private static List<KafkaStream<byte[], byte[]>> kafkaStreams;
@Before
@Override
public void setUp() throws Exception {
super.setUp();
KafkaTestUtil.startZookeeper();
KafkaTestUtil.startKafkaBrokers(3);
KafkaTestUtil.createTopic(TOPIC, 3, 2);
kafkaStreams = KafkaTestUtil.createKafkaStream(KafkaTestUtil.getZkServer().connectString(), TOPIC, 3);
}
@After
@Override
public void tearDown() {
KafkaTestUtil.shutdown();
}
@Override
protected String getPipelineJson() throws Exception {
URI uri = Resources.getResource("kafka_destination_pipeline_run.json").toURI();
String pipelineJson = new String(Files.readAllBytes(Paths.get(uri)), StandardCharsets.UTF_8);
pipelineJson = pipelineJson.replace("topicName", TOPIC);
pipelineJson = pipelineJson.replaceAll("localhost:9092", KafkaTestUtil.getMetadataBrokerURI());
pipelineJson = pipelineJson.replaceAll("localhost:2181", KafkaTestUtil.getZkServer().connectString());
return pipelineJson;
}
@Override
protected int getRecordsInOrigin() {
return 500;
}
@Override
protected int getRecordsInTarget() {
int expectedRecordsInTarget = 0;
for(KafkaStream<byte[], byte[]> kafkaStream : kafkaStreams) {
ConsumerIterator<byte[], byte[]> it = kafkaStream.iterator();
try {
while (it.hasNext()) {
expectedRecordsInTarget++;
it.next();
}
} catch (kafka.consumer.ConsumerTimeoutException e) {
//no-op
}
}
return expectedRecordsInTarget;
}
@Override
protected String getPipelineName() {
return "kafka_destination_pipeline";
}
@Override
protected String getPipelineRev() {
return "0";
}
}