Re: Batch reading from Cassandra

2020-02-28 Thread Piotr Nowojski
Hi, I’m afraid that we don’t have any native support for reading from Cassandra at the moment. The only things that I could find, are streaming sinks [1][2]. Piotrek [1] https://ci.apache.org/projects/flink/flink-docs-stable/dev/connectors/cassandra.html

Batch reading from Cassandra

2020-02-23 Thread Lasse Nedergaard
Hi. We would like to do some batch analytics on our data set stored in Cassandra and are looking for an efficient way to load data from a single table. Not by key, but random 15%, 50% or 100% Data bricks has create an efficient way to load Cassandra data into Apache Spark and they are doing it

Re: Batch reading from Cassandra. How to?

2020-02-17 Thread Till Rohrmann
Hi Lasse, as far as I know, the best way to read from Cassandra is to use the CassandraInputFormat [1]. Unfortunately, there is no such optimized way to read a large amount of data as Spark offers it at the moment. But if you want to contribute this feature to Flink, then the community would highl

Re: Batch reading from Cassandra. How to?

2020-02-14 Thread Lasse Nedergaard
Any good suggestions? Lasse Den tir. 11. feb. 2020 kl. 08.48 skrev Lasse Nedergaard < lassenederga...@gmail.com>: > Hi. > > We would like to do some batch analytics on our data set stored in > Cassandra and are looking for an efficient way to load data from a single > table. Not by key, but rand

Batch reading from Cassandra. How to?

2020-02-10 Thread Lasse Nedergaard
Hi. We would like to do some batch analytics on our data set stored in Cassandra and are looking for an efficient way to load data from a single table. Not by key, but random 15%, 50% or 100% Data bricks has create an efficient way to load Cassandra data into Apache Spark and they are doing it by