This was a regression from the Thrift 0.5 upgrade.  Should be fixed in r1026415

On Fri, Oct 22, 2010 at 5:11 AM, ruslan usifov <ruslan.usi...@gmail.com> wrote:
> I try play with cassandra 0.7 (i build it from trunk) and its looks better
> then 0.6 brunch, but when i try to add new node with auto_bootstrap: true i
> got NPE (192.168.0.37 initial node with data on it, 192.168.0.220
> bootstraped node):
>
> DEBUG 14:00:58,931 Checking to see if compaction of Schema would be useful
> DEBUG 14:00:58,948 Checking to see if compaction of IndexInfo would be
> useful
>  INFO 14:00:58,929 Upgrading to 0.7. Purging hints if there are any. Old
> hints will be snapshotted.
>  INFO 14:00:58,954 Cassandra version: 0.7.0-beta2-SNAPSHOT
>  INFO 14:00:58,954 Thrift API version: 19.2.0
>  INFO 14:00:58,961 Loading persisted ring state
>  INFO 14:00:58,962 Starting up server gossip
>  INFO 14:00:58,968 switching in a fresh Memtable for LocationInfo at
> CommitLogContext(file='/data/cassandra/0.7/commitlog/CommitLog-1
> 287741658826.log', position=700)
>  INFO 14:00:58,969 Enqueuing flush of memtable-locationi...@14222419(227
> bytes, 4 operations)
>  INFO 14:00:58,970 Writing memtable-locationi...@14222419(227 bytes, 4
> operations)
>  INFO 14:00:59,089 Completed flushing
> /data/cassandra/0.7/data/system/LocationInfo-e-1-Data.db
> DEBUG 14:00:59,093 Checking to see if compaction of LocationInfo would be
> useful
> DEBUG 14:00:59,094 discard completed log segments for
> CommitLogContext(file='/data/cassandra/0.7/commitlog/CommitLog-1287741658826.lo
> g', position=700), column family 0.
> DEBUG 14:00:59,095 Marking replay position 700 on commit log
> CommitLogSegment(/data/cassandra/0.7/commitlog/CommitLog-1287741658826.l
> og)
> DEBUG 14:00:59,116 attempting to connect to /192.168.0.37
> ERROR 14:00:59,118 Exception encountered during startup.
> java.lang.NullPointerException
>         at
> org.apache.cassandra.db.SystemTable.isBootstrapped(SystemTable.java:308)
>         at
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:437)
>         at
> org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:159)
>         at
> org.apache.cassandra.thrift.CassandraDaemon.setup(CassandraDaemon.java:55)
>         at
> org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCassandraDaemon.java:215)
>         at
> org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:134)
> Exception encountered during startup.
> java.lang.NullPointerException
>         at
> org.apache.cassandra.db.SystemTable.isBootstrapped(SystemTable.java:308)
>         at
> org.apache.cassandra.service.StorageService.initServer(StorageService.java:437)
>         at
> org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:159)
>         at
> org.apache.cassandra.thrift.CassandraDaemon.setup(CassandraDaemon.java:55)
>         at
> org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCassandraDaemon.java:215)
>         at
> org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:134)
>
>
>
> Is it bug or i do something wrong?
>
>
>
>
>
>
> PS: here is my cassandra yaml
>
> # Cassandra storage config YAML
>
> cluster_name: 'Test Cluster'
>
> initial_token:
>
> auto_bootstrap: true
>
> hinted_handoff_enabled: true
>
> authenticator: org.apache.cassandra.auth.AllowAllAuthenticator
>
> authority: org.apache.cassandra.auth.AllowAllAuthority
>
> partitioner: org.apache.cassandra.dht.RandomPartitioner
>
> # directories where Cassandra should store data on disk.
> data_file_directories:
>     - /data/cassandra/0.7/data
>
> # commit log
> commitlog_directory: /data/cassandra/0.7/commitlog
>
> # saved caches
> saved_caches_directory: /data/cassandra/0.7/saved_caches
>
> # Size to allow commitlog to grow to before creating a new segment
> commitlog_rotation_threshold_in_mb: 128
>
> commitlog_sync: periodic
>
> commitlog_sync_period_in_ms: 10000
>
> seeds:
>     - 192.168.0.37
>
> disk_access_mode: auto
>
> concurrent_reads: 8
> concurrent_writes: 32
>
> memtable_flush_writers: 1
>
> # TCP port, for commands and data
> storage_port: 7000
> listen_address: 192.168.0.220
>
> rpc_address: 192.168.0.220
> rpc_port: 9160
>
> # enable or disable keepalive on rpc connections
> rpc_keepalive: true
>
> binary_memtable_throughput_in_mb: 256
>
> # Add column indexes to a row after its contents reach this size.
> # Increase if your column values are large, or if you have a very large
> # number of columns.  The competing causes are, Cassandra has to
> # deserialize this much of the row to read a single column, so you want
> # it to be small - at least if you do many partial-row reads - but all
> # the index data is read for each access, so you don't want to generate
> # that wastefully either.
> column_index_size_in_kb: 64
>
> # Size limit for rows being compacted in memory.  Larger rows will spill
> # over to disk and use a slower two-pass compaction process.  A message
> # will be logged specifying the row key.
> in_memory_compaction_limit_in_mb: 64
>
> # Time to wait for a reply from other nodes before failing the command
> rpc_timeout_in_ms: 10000
>
> # phi value that must be reached for a host to be marked down.
> # most users should never need to adjust this.
> # phi_convict_threshold: 8
>
> # endpoint_snitch -- Set this to a class that implements
> # IEndpointSnitch, which will let Cassandra know enough
> # about your network topology to route requests efficiently.
> # Out of the box, Cassandra provides
> #  - org.apache.cassandra.locator.SimpleSnitch:
> #    Treats Strategy order as proximity. This improves cache locality
> #    when disabling read repair, which can further improve throughput.
> #  - org.apache.cassandra.locator.RackInferringSnitch:
> #    Proximity is determined by rack and data center, which are
> #    assumed to correspond to the 3rd and 2nd octet of each node's
> #    IP address, respectively
> # org.apache.cassandra.locator.PropertyFileSnitch:
> #  - Proximity is determined by rack and data center, which are
> #    explicitly configured in cassandra-rack.properties.
> endpoint_snitch: org.apache.cassandra.locator.SimpleSnitch
>
> # dynamic_snitch -- This boolean controls whether the above snitch is
> # wrapped with a dynamic snitch, which will monitor read latencies
> # and avoid reading from hosts that have slowed (due to compaction,
> # for instance)
> dynamic_snitch: true
> # controls how often to perform the more expensive part of host score
> # calculation
> dynamic_snitch_update_interval_in_ms: 100
> # controls how often to reset all host scores, allowing a bad host to
> # possibly recover
> dynamic_snitch_reset_interval_in_ms: 600000
> # if set greater than zero and read_repair_chance is < 1.0, this will allow
> # 'pinning' of replicas to hosts in order to increase cache capacity.
> # The badness threshold will control how much worse the pinned host has to
> be
> # before the dynamic snitch will prefer other replicas over it.  This is
> # expressed as a double which represents a percentage.
> dynamic_snitch_badness_threshold: 0.0
>
> # request_scheduler -- Set this to a class that implements
> # RequestScheduler, which will schedule incoming client requests
> # according to the specific policy. This is useful for multi-tenancy
> # with a single Cassandra cluster.
> # NOTE: This is specifically for requests from the client and does
> # not affect inter node communication.
> # org.apache.cassandra.scheduler.NoScheduler - No scheduling takes place
> # org.apache.cassandra.scheduler.RoundRobinScheduler - Round robin of
> # client requests to a node with a separate queue for each
> # request_scheduler_id. The scheduler is further customized by
> # request_scheduler_options as described below.
> request_scheduler: org.apache.cassandra.scheduler.NoScheduler
>
> # Scheduler Options vary based on the type of scheduler
> # NoScheduler - Has no options
> # RoundRobin
> #  - throttle_limit -- The throttle_limit is the number of in-flight
> #                      requests per client.  Requests beyond
> #                      that limit are queued up until
> #                      running requests can complete.
> #                      The value of 80 here is twice the number of
> #                      concurrent_reads + concurrent_writes.
> #  - default_weight -- default_weight is optional and allows for
> #                      overriding the default which is 1.
> #  - weights -- Weights are optional and will default to 1 or the
> #               overridden default_weight. The weight translates into how
> #               many requests are handled during each turn of the
> #               RoundRobin, based on the scheduler id.
> #
> # request_scheduler_options:
> #    throttle_limit: 80
> #    default_weight: 5
> #    weights:
> #      Keyspace1: 1
> #      Keyspace2: 5
>
> # request_scheduler_id -- An identifer based on which to perform
> # the request scheduling. Currently the only valid option is keyspace.
> # request_scheduler_id: keyspace
>
> # The Index Interval determines how large the sampling of row keys
> #  is for a given SSTable. The larger the sampling, the more effective
> #  the index is at the cost of space.
> index_interval: 128
>
>
>



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of Riptano, the source for professional Cassandra support
http://riptano.com

Reply via email to