Hi Yonghui,

I have tried a few ways with different retention strategies to try and reproduce this issue, but haven't been able to do it. Since it looks like you can consistently reproduce this, would you able to share a sample reproducible application (maybe as a github repo) for this issue?

-Jaikiran
On Monday 26 January 2015 09:27 AM, Yonghui Zhao wrote:
I have fixed this issue like this patch
https://reviews.apache.org/r/29755/diff/5/.

I find rename failure still happens:

server.log.2015-01-26-06:[2015-01-26 06:10:54,513] ERROR File rename
failed, forcefully deleting file (kafka.log.Log)
server.log.2015-01-26-06:[2015-01-26 06:10:54,600] ERROR File rename
failed, forcefully deleting file (kafka.log.Log)
server.log.2015-01-26-06:[2015-01-26 06:10:54,685] ERROR File rename
failed, forcefully deleting file (kafka.log.Log)
server.log.2015-01-26-06:[2015-01-26 06:10:54,797] ERROR File rename
failed, forcefully deleting file (kafka.log.Log)
....

And use lsof I can still find some files opened by kafka but deleted,  but
those files sizes are 0.

java       3228 root   34uw     REG              253,2         0   26084228
/home/work/data/soft/kafka-0.8/data/.lock (deleted)
java       3228 root   35u      REG              253,2         0   26084232
/home/work/data/soft/kafka-0.8/data/cube-0/00000000000000000000.log
(deleted)
java       3228 root   36u      REG              253,2         0   26869778
/home/work/data/soft/kafka-0.8/data/_oakbay_v2_search_topic_misearch_appstore-search-0/00000000000000003116.log
(deleted)
java       3228 root   37u      REG              253,2         0   26084234
/home/work/data/soft/kafka-0.8/data/_oakbay_v2_search_topic_mishop-search_mishop_v1-0/00000000000000000000.log
(deleted)



Here is my configuration:

Binary: kafka_2.10-0.8.1.1
Retention config:

# The minimum age of a log file to be eligible for deletion
log.retention.hours=168

# A size-based retention policy for logs. Segments are pruned from the log
as long as the remaining
# segments don't drop below log.retention.bytes.
#log.retention.bytes=1073741824

# The maximum size of a log segment file. When this size is reached a new
log segment will be created.
log.segment.bytes=536870912

# The interval at which log segments are checked to see if they can be
deleted according
# to the retention policies
log.retention.check.interval.ms=60000

# By default the log cleaner is disabled and the log retention policy will
default to just delete segments after their retention expires.
# If log.cleaner.enable=true is set the cleaner will be enabled and
individual logs can then be marked for log compaction.
log.cleaner.enable=false



OS:  CentOS release 6.4 (Final)
JDK:
*java version "1.6.0_37"*
Java(TM) SE Runtime Environment (build 1.6.0_37-b06)
Java HotSpot(TM) 64-Bit Server VM (build 20.12-b01, mixed mode)

JDK is too old, but not sure if this result in rename failure.






2015-01-26 0:42 GMT+08:00 Jay Kreps <jay.kr...@gmail.com>:

Also, what is the configuration for the servers? In particular it would be
good to know the retention and/or log compaction settings as those delete
files.

-Jay

On Sun, Jan 25, 2015 at 4:34 AM, Jaikiran Pai <jai.forums2...@gmail.com>
wrote:

Hi Yonghui,

Do you still have this happening? If yes, can you tell us a bit more
about your setup? Is there something else that accesses or maybe deleting
these log files? For more context to this question, please read the
discussion related to this here http://mail-archives.apache.
org/mod_mbox/kafka-dev/201501.mbox/%3C54C47E9B.5060401%40gmail.com%3E


-Jaikiran


On Thursday 08 January 2015 11:19 AM, Yonghui Zhao wrote:

CentOS release 6.3 (Final)


2015-01-07 22:18 GMT+08:00 Harsha <ka...@harsha.io>:

  Yonghui,
             Which OS you are running.
-Harsha

On Wed, Jan 7, 2015, at 01:38 AM, Yonghui Zhao wrote:

Yes  and I found the reason rename in deletion is failed.
In rename progress the files is deleted? and then exception blocks
file
closed in kafka.
But I don't know how can rename failure happen,

[2015-01-07 00:10:48,685] ERROR Uncaught exception in scheduled task
'kafka-log-retention' (kafka.utils.KafkaScheduler)
kafka.common.KafkaStorageException: Failed to change the log file
suffix
from  to .deleted for log segment 70781650
          at kafka.log.LogSegment.changeFileSuffixes(LogSegment.
scala:249)
          at kafka.log.Log.kafka$log$Log$$
asyncDeleteSegment(Log.scala:636)
          at kafka.log.Log.kafka$log$Log$$deleteSegment(Log.scala:627)
          at
kafka.log.Log$$anonfun$deleteOldSegments$1.apply(Log.scala:415)
          at
kafka.log.Log$$anonfun$deleteOldSegments$1.apply(Log.scala:415)
          at scala.collection.immutable.List.foreach(List.scala:318)
          at kafka.log.Log.deleteOldSegments(Log.scala:415)
          at

  
kafka.log.LogManager.kafka$log$LogManager$$cleanupExpiredSegments(LogManager.scala:325)

          at
kafka.log.LogManager$$anonfun$cleanupLogs$3.apply(LogManager.scala:356)

          at
kafka.log.LogManager$$anonfun$cleanupLogs$3.apply(LogManager.scala:354)

          at

  scala.collection.TraversableLike$WithFilter$$
anonfun$foreach$1.apply(TraversableLike.scala:772)

          at scala.collection.Iterator$class.foreach(Iterator.scala:
727)
          at scala.collection.AbstractIterator.foreach(
Iterator.scala:1157)
          at
scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
          at scala.collection.AbstractIterable.foreach(
Iterable.scala:54)
          at

  scala.collection.TraversableLike$WithFilter.
foreach(TraversableLike.scala:771)

          at kafka.log.LogManager.cleanupLogs(LogManager.scala:354)
          at

  kafka.log.LogManager$$anonfun$startup$1.apply$mcV$sp(LogManager.scala:141)

          at
kafka.utils.KafkaScheduler$$anon$1.run(KafkaScheduler.scala:100)
          at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)

          at

  java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java:317)

          at
java.util.concurrent.FutureTask.runAndReset(FutureTask.java:150)
          at

  java.util.concurrent.ScheduledThreadPoolExecutor$
ScheduledFutureTask.access$101(ScheduledThreadPoolExecutor.java:98)

          at

  java.util.concurrent.ScheduledThreadPoolExecutor$
ScheduledFutureTask.runPeriodic(ScheduledThreadPoolExecutor.java:180)

          at

  java.util.concurrent.ScheduledThreadPoolExecutor$
ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:204)

          at

  java.util.concurrent.ThreadPoolExecutor$Worker.
runTask(ThreadPoolExecutor.java:886)

          at

  
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)

          at java.lang.Thread.run(Thread.java:662)


2015-01-07 13:56 GMT+08:00 Jun Rao <j...@confluent.io>:

  Do you mean that the Kafka broker still holds a file handler on a
deleted
file? Do you see those files being deleted in the Kafka log4j log?
Thanks,

Jun

On Tue, Jan 6, 2015 at 4:46 AM, Yonghui Zhao <zhaoyong...@gmail.com>
wrote:

  Hi,
We use kafka_2.10-0.8.1.1 in our server. Today we found disk space

alert.
We find many kafka data files are deleted, but still opened by kafka.
such as:

_yellowpageV2-0/00000000000068170670.log (deleted)
java       8446         root  724u      REG 253,2

536937911
26087362

  /home/work/data/soft/kafka-0.8/data/_oakbay_v2_search_
topic_ypgsearch_yellowpageV2-0/00000000000068818668.log

(deleted)
java       8446         root  725u      REG 253,2

536910838
26087364

  /home/work/data/soft/kafka-0.8/data/_oakbay_v2_search_
topic_ypgsearch_yellowpageV2-0/00000000000069457098.log

(deleted)
java       8446         root  726u      REG 253,2

536917902
26087368

  /home/work/data/soft/kafka-0.8/data/_oakbay_v2_search_
topic_ypgsearch_yellowpageV2-0/00000000000070104914.log

(deleted)

Is there anything wrong or wrong configed?



Reply via email to