Cassandra doesn't do that automatically - it can guarantee consistency on read 
or write via ConsistencyLevel on each query, and it can run active 
(AntiEntropy) repairs. But active repairs must be scheduled (by human or cron 
or by third party script like http://cassandra-reaper.io/), and to be pedantic, 
repair only fixes consistency issue, there's some work to be done to properly 
address/support fixing corrupted replicas (for example, repair COULD send a bit 
flip from one node to all of the others)



-- 
Jeff Jirsa


> On Sep 9, 2017, at 1:07 AM, Ralph Soika <ralph.so...@imixs.com> wrote:
> 
> Hi,
> 
> I am searching for a big data storage solution for the Imixs-Workflow 
> project. I started with Hadoop until I became aware of the 
> 'small-file-problem'. So I am considering using Cassandra now. 
> But Hadoop has one important feature for me. The replicator continuously 
> examines whether data blocks are consistent across all datanodes. This will 
> detect disk errors and automatically move data from defective blocks to 
> working blocks. I think this is called 'self-healing mechanism'.
> Is there a similar feature in Cassandra too?
> 
> Thanks for help
> 
> Ralph
> 
> 
> 
> -- 
> 

Reply via email to