Kafka Connect耗尽堆空间。已经设置`-Xmx12g`

时间:2018-03-13 20:51:20

标签: apache-kafka out-of-memory apache-kafka-connect confluent-kafka

我的Kafka Connect接收器的堆空间不足。还有其他类似的主题:Kafka Connect running out of heap space

问题是使用默认内存设置运行。以前,将它提高到2g可以解决我的问题。但是,添加新接收器时,堆错误会恢复。我把Xmx提高到12g,我仍然得到错误。

在我的systemd服务文件中,我有:

Environment="KAFKA_HEAP_OPTS=-Xms512m -Xmx12g"

即使Xmx设置非常高,我仍然会收到堆错误。我还将我的flush.size降低到1000,我认为这会有所帮助。仅供参考,此连接器针对11种不同的Kafka主题。这是否会产生独特的记忆需求?

如何进一步修复或诊断?

仅供参考,这是Kafka 0.10.2.1和Confluent Platform 3.2.2。更新的版本是否提供了任何改进?

java.lang.OutOfMemoryError: Java heap space
at java.nio.HeapByteBuffer.<init>(HeapByteBuffer.java:57)
at java.nio.ByteBuffer.allocate(ByteBuffer.java:335)
at io.confluent.connect.s3.storage.S3OutputStream.<init>(S3OutputStream.java:67)
at io.confluent.connect.s3.storage.S3Storage.create(S3Storage.java:197)
at io.confluent.connect.s3.format.avro.AvroRecordWriterProvider$1.write(AvroRecordWriterProvider.java:67)
at io.confluent.connect.s3.TopicPartitionWriter.writeRecord(TopicPartitionWriter.java:393)
at io.confluent.connect.s3.TopicPartitionWriter.write(TopicPartitionWriter.java:197)
at io.confluent.connect.s3.S3SinkTask.put(S3SinkTask.java:173)
at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:429)
at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:250)
at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:179)
at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:148)
at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:139)
at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:182)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[2018-03-13 20:31:46,398] ERROR Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerSinkTask:450)
[2018-03-13 20:31:46,401] ERROR Task avro-s3-sink-0 threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask:141)
org.apache.kafka.connect.errors.ConnectException: Exiting WorkerSinkTask due to unrecoverable exception.
at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:451)
at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:250)
at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:179)
at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:148)
at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:139)
at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:182)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

2 个答案:

答案 0 :(得分:1)

目前,S3连接器的内存要求取决于未完成分区的数量和prop。尝试将后者设置为5MB(允许的最小值)。默认值为25MB。

另请阅读此处,有关尺寸建议的更详细说明:

source

答案 1 :(得分:0)

首先,我对卡夫卡一无所知。

但是,作为一般规则,当某个流程遇到某种容量限制,并且您无法提高该限制时,您必须以某种方式限制该流程。建议您探索定期暂停的可能性。也许睡眠10毫秒非常100毫秒。这样的事情。

您可以尝试的另一件事是将您的Kafka进程 pin 发送到一个特定的CPU。这有时会产生惊人的有益效果。