> During one of my tests - see this thread in this mailing list: > http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/java-io-IOException-FAILED-TO-UNCOMPRESS-5-exception-when-running-nodetool-rebuild-td7586494.html That thread has been updated, check the bug ondrej created.
> How will this perform in production with much bigger data if repair > takes 25 minutes on 7MB and 11k compactions were triggered by the > repair run? Seems a little odd. See what happens the next time you run repair. Cheers ----------------- Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton http://www.thelastpickle.com On 27/03/2013, at 2:36 AM, Ondřej Černoš <cern...@gmail.com> wrote: > Hi all, > > I have 2 DCs, 3 nodes each, RF:3, I use local quorum for both reads and > writes. > > Currently I test various operational qualities of the setup. > > During one of my tests - see this thread in this mailing list: > http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/java-io-IOException-FAILED-TO-UNCOMPRESS-5-exception-when-running-nodetool-rebuild-td7586494.html > - I ran into this situation: > > - all nodes have all data and agree on it: > > [user@host1-dc1:~] nodetool status > > Datacenter: na-prod > =================== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns > (effective) Host ID Rack > UN XXX.XXX.XXX.XXX 7.74 MB 256 100.0% > 0b1f1d79-52af-4d1b-a86d-bf4b65a05c49 cmp17 > UN XXX.XXX.XXX.XXX 7.74 MB 256 100.0% > 039f206e-da22-44b5-83bd-2513f96ddeac cmp10 > UN XXX.XXX.XXX.XXX 7.72 MB 256 100.0% > 007097e9-17e6-43f7-8dfc-37b082a784c4 cmp11 > Datacenter: us-east > =================== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns > (effective) Host ID Rack > UN XXX.XXX.XXX.XXX 7.73 MB 256 100.0% > a336efae-8d9c-4562-8e2a-b766b479ecb4 1d > UN XXX.XXX.XXX.XXX 7.73 MB 256 100.0% > ab1bbf0a-8ddc-4a12-a925-b119bd2de98e 1d > UN XXX.XXX.XXX.XXX 7.73 MB 256 100.0% > f53fd294-16cc-497e-9613-347f07ac3850 1d > > - only one node disagrees: > > [user@host1-dc2:~] nodetool status > Datacenter: us-east > =================== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns Host ID > Rack > UN XXX.XXX.XXX.XXX 7.73 MB 256 17.6% > a336efae-8d9c-4562-8e2a-b766b479ecb4 1d > UN XXX.XXX.XXX.XXX 7.75 MB 256 16.4% > ab1bbf0a-8ddc-4a12-a925-b119bd2de98e 1d > UN XXX.XXX.XXX.XXX 7.73 MB 256 15.7% > f53fd294-16cc-497e-9613-347f07ac3850 1d > Datacenter: na-prod > =================== > Status=Up/Down > |/ State=Normal/Leaving/Joining/Moving > -- Address Load Tokens Owns Host ID > Rack > UN XXX.XXX.XXX.XXX 7.74 MB 256 16.9% > 0b1f1d79-52af-4d1b-a86d-bf4b65a05c49 cmp17 > UN XXX.XXX.XXX.XXX 7.72 MB 256 17.1% > 007097e9-17e6-43f7-8dfc-37b082a784c4 cmp11 > UN XXX.XXX.XXX.XXX 7.73 MB 256 16.3% > 039f206e-da22-44b5-83bd-2513f96ddeac cmp10 > > I tried to rebuild the node from scratch, repair the node, no results. > Still the same owns stats. > > The cluster is built from cassandra 1.2.3 and uses vnodes. > > > On the related note: the data size, as you can see, is really small. > The cluster was created by setting up the us-east datacenter, > populating it with the dataset, then building the na-prod datacenter > and running nodetool rebuild us-east. When I tried to run nodetool > repair it took 25 minutes to finish, on this small dataset. Is this > ok? > > One other think I notices is the amount of compactions on the system keyspace: > > /.../system/schema_columnfamilies/system-schema_columnfamilies-ib-11694-TOC.txt > /.../system/schema_columnfamilies/system-schema_columnfamilies-ib-11693-Statistics.db > > This is just after running the repair. Is this ok, considering the > dataset is 7MB and during the repair no operations were running > against the database, neither read, nor write, nothing? > > How will this perform in production with much bigger data if repair > takes 25 minutes on 7MB and 11k compactions were triggered by the > repair run? > > regards, > > Ondrej Cernos