Hello people, I'm running one single kafka broker from within a docker container. The folder where kafka writes the logs is mounted as *docker volume* on my system.
As soon as I try to create a topic I get this error [2016-06-15 10:22:53,602] ERROR [KafkaApi-0] Error when handling request {controller_id=0,controller_epoch=1,partition_states=[{topic=mytopic,partition=0,controller_epoch=1,leader=0,leader_epoch=0,isr=[0],zk_version=0,replicas=[0]}],live_leaders=[{id=0,host=kafkadocker,port=9092}]} (kafka.server.KafkaApis) *java.io.IOException: Invalid argument* at sun.nio.ch.FileChannelImpl.map0(Native Method) at sun.nio.ch.FileChannelImpl.map(FileChannelImpl.java:926) *at kafka.log.OffsetIndex.<init>(OffsetIndex.scala:75)* at kafka.log.LogSegment.<init>(LogSegment.scala:58) at kafka.log.Log.loadSegments(Log.scala:233) at kafka.log.Log.<init>(Log.scala:101) at kafka.log.LogManager.createLog(LogManager.scala:363) at kafka.cluster.Partition.getOrCreateReplica(Partition.scala:96) at kafka.cluster.Partition$$anonfun$4$$anonfun$apply$2.apply(Partition.scala:176) at kafka.cluster.Partition$$anonfun$4$$anonfun$apply$2.apply(Partition.scala:176) at scala.collection.mutable.HashSet.foreach(HashSet.scala:78) at kafka.cluster.Partition$$anonfun$4.apply(Partition.scala:176) at kafka.cluster.Partition$$anonfun$4.apply(Partition.scala:170) at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:231) at kafka.utils.CoreUtils$.inWriteLock(CoreUtils.scala:239) at kafka.cluster.Partition.makeLeader(Partition.scala:170) at kafka.server.ReplicaManager$$anonfun$makeLeaders$4.apply(ReplicaManager.scala:699) at kafka.server.ReplicaManager$$anonfun$makeLeaders$4.apply(ReplicaManager.scala:698) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99) at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) at scala.collection.mutable.HashMap.foreach(HashMap.scala:99) at kafka.server.ReplicaManager.makeLeaders(ReplicaManager.scala:698) at kafka.server.ReplicaManager.becomeLeaderOrFollower(ReplicaManager.scala:644) at kafka.server.KafkaApis.handleLeaderAndIsrRequest(KafkaApis.scala:144) at kafka.server.KafkaApis.handle(KafkaApis.scala:80) at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:60) at java.lang.Thread.run(Thread.java:745) The error is an IOException, so it looks like the broker has trouble trying to access the log file. Looks like Kafka assumes a feature of the underlying filesystem, which is not present. I do not get any error if I keep the kafka log-files inside the docker container. Have you seen the issue before ? Thanks. *Valerio*