Kafka无法删除Windows上的旧日志段

时间:2017-05-30 16:43:47

标签: apache-kafka

我在Windows上遇到了Kafka的问题,它试图删除日志段,但由于另一个进程可以访问这些文件,因此无法解决。这是由Kafka保持对文件本身的访问并尝试删除已打开的文件引起的。该错误在下面供参考。

我发现了两个针对此问题https://issues.apache.org/jira/browse/KAFKA-1194https://issues.apache.org/jira/browse/KAFKA-2170提交的JIRA错误。第一个是在0.8.1版本下记录的,第二个是在版本0.10.1下记录的。

我亲自尝试过版本0.10.1和0.10.2。他们都没有修复错误。

我的问题是,有没有人知道可以解决这个问题的补丁,或者知道卡夫卡人是否已经解决了这个问题,很快就会推出。

感谢。

kafka.common.KafkaStorageException: Failed to change the log file suffix from  to .deleted for log segment 6711351
    at kafka.log.LogSegment.kafkaStorageException$1(LogSegment.scala:340)
    at kafka.log.LogSegment.changeFileSuffixes(LogSegment.scala:342)
    at kafka.log.Log.kafka$log$Log$$asyncDeleteSegment(Log.scala:981)
    at kafka.log.Log.kafka$log$Log$$deleteSegment(Log.scala:971)
    at kafka.log.Log$$anonfun$deleteOldSegments$1.apply(Log.scala:673)
    at kafka.log.Log$$anonfun$deleteOldSegments$1.apply(Log.scala:673)
    at scala.collection.immutable.List.foreach(List.scala:381)
    at kafka.log.Log.deleteOldSegments(Log.scala:673)
    at kafka.log.Log.deleteRetentionSizeBreachedSegments(Log.scala:717)
    at kafka.log.Log.deleteOldSegments(Log.scala:697)
    at kafka.log.LogManager$$anonfun$cleanupLogs$3.apply(LogManager.scala:474)
    at kafka.log.LogManager$$anonfun$cleanupLogs$3.apply(LogManager.scala:472)
    at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
    at scala.collection.Iterator$class.foreach(Iterator.scala:893)
    at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
    at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
    at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
    at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
    at kafka.log.LogManager.cleanupLogs(LogManager.scala:472)
    at kafka.log.LogManager$$anonfun$startup$1.apply$mcV$sp(LogManager.scala:200)
    at kafka.utils.KafkaScheduler$$anonfun$1.apply$mcV$sp(KafkaScheduler.scala:110)
    at kafka.utils.CoreUtils$$anon$1.run(CoreUtils.scala:57)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:748)
Caused by: java.nio.file.FileSystemException: c:\kafka-logs\kafka-logs\metric-values-0\00000000000006711351.log -> c:\kafka-logs\kafka-logs\metric-values-0\00000000000006711351.log.deleted: The process cannot access the file because it is being used by another process.

    at sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86)
    at sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97)
    at sun.nio.fs.WindowsFileCopy.move(WindowsFileCopy.java:387)
    at sun.nio.fs.WindowsFileSystemProvider.move(WindowsFileSystemProvider.java:287)
    at java.nio.file.Files.move(Files.java:1395)
    at org.apache.kafka.common.utils.Utils.atomicMoveWithFallback(Utils.java:711)
    at org.apache.kafka.common.record.FileRecords.renameTo(FileRecords.java:210)
    ... 28 more
    Suppressed: java.nio.file.FileSystemException: c:\kafka-logs\kafka-logs\metric-values-0\00000000000006711351.log -> c:\kafka-logs\kafka-logs\metric-values-0\00000000000006711351.log.deleted: The process cannot access the file because it is being used by another process.

            at sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86)
            at sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97)
            at sun.nio.fs.WindowsFileCopy.move(WindowsFileCopy.java:301)
            at sun.nio.fs.WindowsFileSystemProvider.move(WindowsFileSystemProvider.java:287)
            at java.nio.file.Files.move(Files.java:1395)
            at org.apache.kafka.common.utils.Utils.atomicMoveWithFallback(Utils.java:708)
            ... 29 more

1 个答案:

答案 0 :(得分:0)

在本地运行kafka时遇到类似的问题,kafka服务器似乎在删除日志文件失败时停止执行。为了避免这种情况的发生,我必须增加日志的日志保留以避免自动删除。

# The minimum age of a log file to be eligible for deletion due to age log.retention.hours=500

将日志设置为xxx小时将避免在本地运行时发生这种情况,但是对于生产环境,我认为对于基于linux的系统,这应该不会发生。

如果需要删除日志文件,请在日志所在的位置手动将其删除,然后重新启动kafka。