Hello, I'm trying to bulk insert data with a Spark job into our Apache Cassandra 3.7. For this I'm trying to use com.datastax.bdp.spark.writer.BulkTableWriter from Datastax.
I managed to set a up a job which creates the sstables using dse-spark 5.1.5 . However, when they are being loaded on the Cassandra nodes, I see several exceptions in the Cassandra logs, like: Streaming error occurred on session with peer xxx java.lang.IllegalArgumentException: Unknown type XX Error while reading compressed input stream. java.nio.channels.ClosedChannelException: null Streaming error occurred on session with peer xxx java.io.IOException: Corrupt (negative) value length encountered I tried out older version from 4.8 without any success. It looks like that there are always protocol / binary incompatibility issues. Is it possible somehow to use dse-spark with an Apache release of Cassandra (v3.x / 3.7) or does this only work for the Datastax Cassandra version? Thanks, Peter