It’s going to be very difficult to diagnose remotely. I don’t run or have an opinion on jdk7 but I would suspect the following:
- bad hardware (dimm, disk, network card, motherboard, processor in the order) - bad jdk7. I’d be inclined to upgrade to 8 personally, but rolling back to previous version may not be a bad idea You’re in a tough spot if this is spreading. I’d personally be looking to try to isolate the source and roll forward or backward as quickly as possible. I don’t really suspect a cassandra 2.1 but here but it’s possible I suppose. Take a snapshot now as you may need it to try to recover data later. -- Jeff Jirsa > On May 9, 2019, at 10:53 PM, keshava <keshava.kous...@gmail.com> wrote: > > yes we do have compression enabled using > "org.apache.cassandra.io.compress.LZ4Compressor" it is spreading.. > as the no of inserts increases it is spreading across. > yes it did started with JDK and OS upgrade. > > Best regards :) > keshava Hosahalli > > >> On Thu, May 9, 2019 at 7:11 PM Jeff Jirsa <jji...@gmail.com> wrote: >> Do you have compression enabled on your table? >> >> Did this start with the JDK upgrade? >> >> Is the compression spreading, or is it contained to the same % of entries? >> >> >> >>> On Thu, May 9, 2019 at 4:12 AM keshava <keshava.kous...@gmail.com> wrote: >>> Hi , our application is running in to data corruption issue. Application >>> uses cassandra 2.1.11 with datastax java driver version 2.1.9. So far all >>> working fine. recently we changed our deployment environment to openjdk >>> 1.7.191 (earlier it was 1.7.181) and centos 7.4 (earlier 6.8) This is >>> randomly happening for one table. 1 in every 4-5 entries are getting >>> corrupted writing new entries will return success and when i try to read i >>> get data not found .whenist all the data available in the table using >>> cqlsh i see garbage entries like >>> >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> >>> here is the output of the cqlsh >>> >>> cqlsh:ccp> select id from socialcontact; >>> >>> id >>> ------------------------------------------------------------------------------>---------------------------------------------------- >>> >>> 9BA31AE31000016A0000097C3F57FEF9 9BA10FB21000016A000000103F57FEF9 >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> 9BA3236C1000016A000009E63F57FEF9 9BA325361000016A000009FC3F57FEF9 >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 >>> >>> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00. >>> >>> I did enabled the query tracing on both cassandra server and driver. didn't >>> noticed any differences. looking for any advice's in resolving this issue >>> >>> PS i did tried upgrading cassandra to latest in 2.1 train but it didn't help >>>