我正在使用Apache Kafka和Avro Serializer,使用特定格式。我正在尝试创建自己的自定义类并用作kafka消息值。但是,当我试图发送消息时,我得到以下异常:
Exception in thread "main" org.apache.avro.AvroRuntimeException: not open
at org.apache.avro.file.DataFileWriter.assertOpen(DataFileWriter.java:82)
at org.apache.avro.file.DataFileWriter.append(DataFileWriter.java:287)
at com.harmeetsingh13.java.producers.avroserializer.AvroProducer.main(AvroProducer.java:57)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:147)
My Avro Schema文件如下:
{
"namespace": "customer.avro",
"type": "record",
"name": "Customer",
"fields": [{
"name": "id",
"type": "int"
}, {
"name": "name",
"type": "string"
}]
}
客户类:
public class Customer {
public int id;
public String name;
public Customer() {
}
public Customer(int id, String name) {
this.id = id;
this.name = name;
}
public int getId() {
return id;
}
public void setId(int id) {
this.id = id;
}
public String getName() {
return name;
}
public void setName(String name) {
this.name = name;
}
}
使用Avro进行数据序列化:
public static void fireAndForget(ProducerRecord<String, DataFileWriter> record) {
kafkaProducer.send(record);
}
Customer customer1 = new Customer(1001, "James");
Parser parser = new Parser();
Schema schema = parser.parse(AvroProducer.class.getClassLoader().getResourceAsStream("customer.avro"));
SpecificDatumWriter<Customer> writer = new SpecificDatumWriter<>(schema);
DataFileWriter<Customer> dataFileWriter = new DataFileWriter<>(writer);
dataFileWriter.append(customer1);
dataFileWriter.close();
ProducerRecord<String, DataFileWriter> record1 = new ProducerRecord<>("CustomerCountry",
"Customer One", dataFileWriter
);
fireAndForget(record1);
我想使用SpecificDatumWriter
编写器而不是通用编写器。这个错误与什么有关?
答案 0 :(得分:1)
Kafka收到一个要序列化的键值对,你传递的是DataFileWriter
,这不是你要序列化的值,但这不会起作用。
您需要做的是通过BinaryEncoder
和ByteArrayOutputStream
创建一个包含序列化avro的字节数组,然后将其传递给ProducerRecord<String, byte[]>
:
SpecificDatumWriter<Customer> writer = new SpecificDatumWriter<>(schema);
ByteArrayOutputStream os = new ByteArrayOutputStream();
try {
BinaryEncoder encoder = EncoderFactory.get().binaryEncoder(os, null);
writer.write(customer1, encoder);
e.flush();
byte[] avroBytes = os.toByteArray();
ProducerRecord<String, byte[]> record1 =
new ProducerRecord<>("CustomerCountry", "Customer One", avroBytes);
kafkaProducer.send(record1);
} finally {
os.close();
}