Hello,

I'm trying to bulk insert data with a Spark job into our Apache Cassandra
3.7.
For this I'm trying to use com.datastax.bdp.spark.writer.BulkTableWriter
from Datastax.

I managed to set a up a job which creates the sstables using dse-spark
5.1.5 .
However, when they are being loaded on the Cassandra nodes, I see several
exceptions in the Cassandra logs, like:

Streaming error occurred on session with peer xxx
java.lang.IllegalArgumentException: Unknown type XX

Error while reading compressed input stream.
java.nio.channels.ClosedChannelException: null

Streaming error occurred on session with peer xxx java.io.IOException:
Corrupt (negative) value length encountered

I tried out older version from 4.8 without any success. It looks like that
there are always protocol / binary incompatibility issues.

Is it possible somehow to use dse-spark with an Apache release of Cassandra
(v3.x / 3.7) or does this only work for the Datastax Cassandra version?


Thanks,
Peter

Reply via email to