Hi, I'd be very interested in how you figured it out. Mind sharing?
Jacek On 18 Jan 2017 9:51 p.m., "mhornbech" <mor...@datasolvr.com> wrote: > For anyone revisiting this at a later point, the issue was that Spark 2.1.0 > upgrades netty to version 4.0.42 which is not binary compatible with > version > 4.0.37 used by version 3.1.0 of the Cassandra Java Driver. The newer > version > can work with Cassandra, but because of differences in the maven artifacts > (Spark depends on netty-all while Cassandra depends on netty-transport) > this > was not automatically resolved by SBT. Adding an explicit dependency to > netty-transport version 4.0.42 solved the problem. > > > > -- > View this message in context: http://apache-spark-user-list. > 1001560.n3.nabble.com/New-runtime-exception-after-switch-to-Spark-2-1-0- > tp28263p28319.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >