Aaron, I updated the cassandra files and but still receive the same error (on client side) with a different line number 551:
org.apache.thrift.TApplicationException: Internal error processing get_slice at org.apache.thrift.TApplicationException.read(TApplicationException.java:108) at org.apache.cassandra.thrift.Cassandra$Client.recv_get_slice(Cassandra.java:551) at org.apache.cassandra.thrift.Cassandra$Client.get_slice(Cassandra.java:531) at org.scale7.cassandra.pelops.Selector$6.execute(Selector.java:538) at org.scale7.cassandra.pelops.Selector$6.execute(Selector.java:535) at org.scale7.cassandra.pelops.Operand.tryOperation(Operand.java:45) at org.scale7.cassandra.pelops.Selector.getSuperColumnsFromRow(Selector.java:545) at org.scale7.cassandra.pelops.Selector.getSuperColumnsFromRow(Selector.java:522) at com.isidorey.cassandra.dao.CassandraDAO.getSuperColumnsByKey(CassandraDAO.java:36) at com.isidorey.cassandra.dao.CassandraDAO.getSuperColumnMap(CassandraDAO.java:82) On the server side, this is what we're seeing in Cassandra's log file: ERROR [pool-1-thread-2486] 2010-10-15 17:15:39,740 Cassandra.java (line 2876) Internal error processing get_slice java.lang.RuntimeException: org.apache.cassandra.db.UnserializableColumnFamilyException: Couldn't find cfId=1052 at org.apache.cassandra.thrift.CassandraServer.readColumnFamily(CassandraServer.java:133) at org.apache.cassandra.thrift.CassandraServer.getSlice(CassandraServer.java:222) at org.apache.cassandra.thrift.CassandraServer.multigetSliceInternal(CassandraServer.java:300) at org.apache.cassandra.thrift.CassandraServer.get_slice(CassandraServer.java:261) at org.apache.cassandra.thrift.Cassandra$Processor$get_slice.process(Cassandra.java:2868) at org.apache.cassandra.thrift.Cassandra$Processor.process(Cassandra.java:2724) at org.apache.cassandra.thrift.CustomTThreadPoolServer$WorkerProcess.run(CustomTThreadPoolServer.java:167) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:636) Caused by: org.apache.cassandra.db.UnserializableColumnFamilyException: Couldn't find cfId=1052 at org.apache.cassandra.db.ColumnFamilySerializer.deserialize(ColumnFamilySerializer.java:113) at org.apache.cassandra.db.RowSerializer.deserialize(Row.java:76) at org.apache.cassandra.db.ReadResponseSerializer.deserialize(ReadResponse.java:114) at org.apache.cassandra.db.ReadResponseSerializer.deserialize(ReadResponse.java:90) at org.apache.cassandra.service.StorageProxy.weakRead(StorageProxy.java:289) at org.apache.cassandra.service.StorageProxy.readProtocol(StorageProxy.java:220) at org.apache.cassandra.thrift.CassandraServer.readColumnFamily(CassandraServer.java:120) On Thu, Oct 14, 2010 at 6:29 PM, Aaron Morton <aa...@thelastpickle.com>wrote: > Am guessing but it looks like cassandra returned an error and the client > then had trouble reading the error. > > However if I look at the Beta 2 java thrift interface in Cassandra.java, > line 544 is not in recv_get_slice. May be nothing. > > Perhaps check the server for an error and double check your client is coded > for beta 2. > > Hope that helps. > > Aaron > > > On 15 Oct, 2010,at 12:32 PM, Frank LoVecchio <fr...@isidorey.com> wrote: > > 10:10:21,787 ERROR ~ Error getting Sensor > > org.apache.thrift.TApplicationException: Internal error processing > get_slice > > at org.apache.thrift.TApplicationException.read( > TApplicationException.java:108) > > at org.apache.cassandra.thrift.Cassandra$Client.recv_get_slice( > Cassandra.java:544) > > at org.apache.cassandra.thrift.Cassandra$Client.get_slice( > Cassandra.java:524) > > at org.scale7.cassandra.pelops.Selector$6.execute(Selector.java:538) > > at org.scale7.cassandra.pelops.Selector$6.execute(Selector.java:535) > > at org.scale7.cassandra.pelops.Operand.tryOperation(Operand.java:45) > > at org.scale7.cassandra.pelops.Selector.getSuperColumnsFromRow( > Selector.java:545) > > at org.scale7.cassandra.pelops.Selector.getSuperColumnsFromRow( > Selector.java:522) > > > Not sure if this is a pelops thing or thrift. > > I spun up a new cluster of 3 nodes a couple of nights ago with the nightly > build 0.7 beta 2's. When I include all 3 nodes in the Pelops Pool, and run > this: > > List<SuperColumn> cols = selector.getSuperColumnsFromRow( > getColFamilyName(), key, > Selector.newColumnsPredicateAll(true, 1000), > ConsistencyLevel.ONE); > > I get the error above. When I create a new Pool with only 1 node, I get > different data back then if I create a new Pool with another single node, > but I don't get the error. Why is this happening? Here is my > cassandra.yaml file, which I haven't modified on the fly: > > # Configuration wiki: > http://wiki.apache.org/cassandra/StorageConfiguration > authority: org.apache.cassandra.auth.AllowAllAuthority > authenticator: org.apache.cassandra.auth.AllowAllAuthenticator > auto_bootstrap: true > binary_memtable_throughput_in_mb: 256 > cluster_name: DevCluster > column_index_size_in_kb: 64 > commitlog_directory: /cassandra/commitlog > saved_caches_directory: /cassandra/saved_caches > commitlog_rotation_threshold_in_mb: 128 > commitlog_sync: periodic > commitlog_sync_period_in_ms: 10000 > concurrent_reads: 8 > concurrent_writes: 50 > data_file_directories: > - /cassandra/data > disk_access_mode: auto > endpoint_snitch: org.apache.cassandra.locator.SimpleSnitch > dynamic_snitch: true > hinted_handoff_enabled: true > in_memory_compaction_limit_in_mb: 256 > keyspaces: > - name: Keyspace > replica_placement_strategy: > org.apache.cassandra.locator.SimpleStrategy > replication_factor: 3 > column_families: > > - name: Sensor > column_type: Super > compare_with: TimeUUIDType > gc_grace_seconds: 864000 > keys_cached: 200000.0 > preload_row_cache: false > read_repair_chance: 1.0 > rows_cached: 0.0 > > listen_address: ec2-internal-ip1 > memtable_flush_after_mins: 60 > memtable_operations_in_millions: 0.3 > memtable_throughput_in_mb: 64 > partitioner: org.apache.cassandra.dht.RandomPartitioner > phi_convict_threshold: 8 > rpc_port: 9160 > rpc_timeout_in_ms: 10000 > seeds: > - ec2-internal-ip1 > - ec2-internal-ip2 > - ec2-internal-ip3 > sliced_buffer_size_in_kb: 64 > snapshot_before_compaction: false > storage_port: 7000 > thrift_framed_transport_size_in_mb: 15 > thrift_max_message_length_in_mb: 16 > initial_token: > > Frank LoVecchio > Software Engineer > Isidorey LLC > > franklovecchio.com > rodsandricers.com > >