Hello,

I'm no longer able to successfully run `sbt unidoc` in branch-2.0, and the 
problem seems to stem from the addition of Kafka 0.10 support. If I remove 
either the Kafka 0.8 or 0.10 projects from the build then unidoc works. If I 
keep both in I get two dozen inexplicable compilation errors as part of the 
unidoc task execution. Here's the first few:

[error] 
/Users/msa/workspace/spark-2.0/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/CachedKafkaConsumer.scala:50:
 value assign is not a member of 
org.apache.kafka.clients.consumer.KafkaConsumer[K,V]
[error]     c.assign(tps)
[error]       ^
[error] 
/Users/msa/workspace/spark-2.0/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/CachedKafkaConsumer.scala:95:
 too many arguments for method seek: (x$1: 
java.util.Map[org.apache.kafka.common.TopicPartition,Long])Unit
[error]     consumer.seek(topicPartition, offset)
[error]                  ^
[error] 
/Users/msa/workspace/spark-2.0/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/CachedKafkaConsumer.scala:100:
 value records is not a member of 
java.util.Map[String,org.apache.kafka.clients.consumer.ConsumerRecords[K,V]]
[error]     val r = p.records(topicPartition)

Running `sbt compile` completes without error.

Has anyone else seen this behavior? Any ideas? This seems to be an issue around 
dependency management, but I'm otherwise stumped.

Cheers,

Michael
---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to