Any ideas, please? To repeat, we are using the exact same cassandra-version on all 4 nodes (2.1.10).
On Fri, Oct 23, 2015 at 9:43 AM, Ajay Garg <ajaygargn...@gmail.com> wrote: > Hi Michael. > > Please find below the contents of cassandra.yaml for CAS11 (the files on > the rest of the three nodes are also exactly the same, except the > "initial_token" and "listen_address" fields) :: > > CAS11 :: > > #################################### > cluster_name: 'InstaMsg Cluster' > num_tokens: 256 > initial_token: -9223372036854775808 > hinted_handoff_enabled: true > max_hint_window_in_ms: 10800000 # 3 hours > hinted_handoff_throttle_in_kb: 1024 > max_hints_delivery_threads: 2 > batchlog_replay_throttle_in_kb: 1024 > authenticator: AllowAllAuthenticator > authorizer: AllowAllAuthorizer > permissions_validity_in_ms: 2000 > partitioner: org.apache.cassandra.dht.Murmur3Partitioner > data_file_directories: > - /var/lib/cassandra/data > > commitlog_directory: /var/lib/cassandra/commitlog > > disk_failure_policy: stop > commit_failure_policy: stop > key_cache_size_in_mb: > key_cache_save_period: 14400 > row_cache_size_in_mb: 0 > row_cache_save_period: 0 > counter_cache_size_in_mb: > counter_cache_save_period: 7200 > saved_caches_directory: /var/lib/cassandra/saved_caches > commitlog_sync: periodic > commitlog_sync_period_in_ms: 10000 > commitlog_segment_size_in_mb: 32 > seed_provider: > - class_name: org.apache.cassandra.locator.SimpleSeedProvider > parameters: > - seeds: "104.239.200.33,119.9.92.77" > > concurrent_reads: 32 > concurrent_writes: 32 > concurrent_counter_writes: 32 > > memtable_allocation_type: heap_buffers > > index_summary_capacity_in_mb: > index_summary_resize_interval_in_minutes: 60 > trickle_fsync: false > trickle_fsync_interval_in_kb: 10240 > storage_port: 7000 > ssl_storage_port: 7001 > listen_address: 104.239.200.33 > start_native_transport: true > native_transport_port: 9042 > start_rpc: true > rpc_address: localhost > rpc_port: 9160 > rpc_keepalive: true > > rpc_server_type: sync > thrift_framed_transport_size_in_mb: 15 > incremental_backups: false > snapshot_before_compaction: false > auto_snapshot: true > > tombstone_warn_threshold: 1000 > tombstone_failure_threshold: 100000 > > column_index_size_in_kb: 64 > batch_size_warn_threshold_in_kb: 5 > > compaction_throughput_mb_per_sec: 16 > compaction_large_partition_warning_threshold_mb: 100 > > sstable_preemptive_open_interval_in_mb: 50 > > read_request_timeout_in_ms: 5000 > range_request_timeout_in_ms: 10000 > > write_request_timeout_in_ms: 2000 > counter_write_request_timeout_in_ms: 5000 > cas_contention_timeout_in_ms: 1000 > truncate_request_timeout_in_ms: 60000 > request_timeout_in_ms: 10000 > cross_node_timeout: false > endpoint_snitch: PropertyFileSnitch > > dynamic_snitch_update_interval_in_ms: 100 > dynamic_snitch_reset_interval_in_ms: 600000 > dynamic_snitch_badness_threshold: 0.1 > > request_scheduler: org.apache.cassandra.scheduler.NoScheduler > > server_encryption_options: > internode_encryption: none > keystore: conf/.keystore > keystore_password: cassandra > truststore: conf/.truststore > truststore_password: cassandra > > client_encryption_options: > enabled: false > keystore: conf/.keystore > keystore_password: cassandra > > internode_compression: all > inter_dc_tcp_nodelay: false > #################################### > > > What changes need to be made, so that whenever a downed server comes back > up, the missing data comes back over to it? > > Thanks and Regards, > Ajay > > > > On Fri, Oct 23, 2015 at 9:05 AM, Michael Shuler <mich...@pbandjelly.org> > wrote: > >> On 10/22/2015 10:14 PM, Ajay Garg wrote: >> >>> However, CAS11 refuses to come up now. >>> Following is the error in /var/log/cassandra/system.log :: >>> >>> >>> ################################################################ >>> ERROR [main] 2015-10-23 03:07:34,242 CassandraDaemon.java:391 - Fatal >>> configuration error >>> org.apache.cassandra.exceptions.ConfigurationException: Cannot change >>> the number of tokens from 1 to 256 >>> >> >> Check your cassandra.yaml - this node has vnodes enabled in the >> configuration when it did not, previously. Check all nodes. Something >> changed. Mixed vnode/non-vnode clusters is bad juju. >> >> -- >> Kind regards, >> Michael >> > > > > -- > Regards, > Ajay > -- Regards, Ajay