/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package gobblin.converter;
import java.io.IOException;
import org.apache.avro.Schema;
import org.apache.avro.generic.GenericRecord;
import com.google.common.base.Preconditions;
import gobblin.configuration.WorkUnitState;
import gobblin.kafka.schemareg.KafkaSchemaRegistry;
import gobblin.kafka.schemareg.KafkaSchemaRegistryFactory;
import gobblin.kafka.schemareg.SchemaRegistryException;
import gobblin.kafka.serialize.LiAvroDeserializerBase;
import gobblin.kafka.serialize.SerializationException;
import gobblin.source.extractor.extract.kafka.KafkaSource;
import gobblin.util.EmptyIterable;
import lombok.extern.slf4j.Slf4j;
/**
* Converts LiKafka byte array messages into avro.
*/
@Slf4j
public class LiKafkaByteArrayMsgToAvroConverter<S> extends ToAvroConverterBase<S, byte[]> {
KafkaSchemaRegistry schemaRegistry;
@Override
public Converter<S, Schema, byte[], GenericRecord> init(WorkUnitState workUnit) {
this.schemaRegistry = KafkaSchemaRegistryFactory.getSchemaRegistry(workUnit.getProperties());
return this;
}
@Override
public Schema convertSchema(S schemaIn, WorkUnitState workUnit)
throws SchemaConversionException {
Preconditions.checkArgument(workUnit.contains(KafkaSource.TOPIC_NAME), "Must specify topic name.");
String topic = workUnit.getProp(KafkaSource.TOPIC_NAME);
try {
return (Schema) this.schemaRegistry.getLatestSchema(topic);
} catch (IOException | SchemaRegistryException e) {
throw new SchemaConversionException(e);
}
}
@Override
public Iterable<GenericRecord> convertRecord(Schema outputSchema, byte[] inputRecord, WorkUnitState workUnit)
throws DataConversionException {
try {
String topic = workUnit.getProp(KafkaSource.TOPIC_NAME);
GenericRecord record = new LiAvroDeserializerBase(schemaRegistry).deserialize(topic, inputRecord);
return new SingleRecordIterable<>(record);
} catch (SerializationException e) {
log.error("Cannot decode one record.", e);
return new EmptyIterable<GenericRecord>();
}
}
}