Re: extending multiget

2010-06-18 Thread Sonny Heer
call? On Thu, Jun 17, 2010 at 5:54 PM, Jonathan Ellis wrote: > No.  At that point you basically have no overhead advantage vs just > doing multiple single-row requests. > > On Thu, Jun 17, 2010 at 2:39 PM, Sonny Heer wrote: >> Any plans for this sort of call? >> >>

extending multiget

2010-06-17 Thread Sonny Heer
Any plans for this sort of call? Instead of: public Map> multiget_slice(String keyspace, List keys, ColumnParent column_parent, SlicePredicate predicate, ConsistencyLevel consistency_level) throws InvalidRequestException, UnavailableException, TimedOutException, TException; --- public

Re: Timeouts running batch_mutate

2010-05-20 Thread Sonny Heer
meant to say OPP :) On Thu, May 20, 2010 at 8:21 AM, Sonny Heer wrote: > Yes, I'm using OOP, because of the way we modeled our data.  Does > Cassandra not handle OOP intensive write operations?  Is HBase a > better approach if one must use OOP? > > > On Thu, May 20, 20

Re: Timeouts running batch_mutate

2010-05-20 Thread Sonny Heer
ike this, > which is why most people deploy on RP. > > If you are using RP you may simply need to add C* capacity, or take > TimeoutException as a signal to throttle your activity. > > On Tue, May 18, 2010 at 4:37 PM, Sonny Heer wrote: >> Yeah there are many writes happening at

Re: Timeouts running batch_mutate

2010-05-18 Thread Sonny Heer
hould be sufficient > > you can turn on debug logging to see how long it's actually taking the > destination node to do the write (or look at cfstats, if no other > writes are going on) > > On Fri, May 14, 2010 at 11:55 AM, Sonny Heer wrote: >> Hey, >> >>

Re: chiton

2010-05-16 Thread Sonny Heer
> On Sun, May 16, 2010 at 1:29 PM, Sonny Heer wrote: > Look like you hit this bug which is now solved: > http://github.com/driftx/chiton/issues/closed#issue/3 sweet, thanks. > It will display the raw value for most column types except UUIDs, which it > will decode.  It does

Re: chiton

2010-05-16 Thread Sonny Heer
rest is a text utf-8)? Thanks On Sun, May 16, 2010 at 11:00 AM, Sonny Heer wrote: > thanks for the help.  I had two versions of python, it was using the wrong > one. > > I got the GUI to show up, but when i connect by going to file/connect > and enter localhost/9160..

Re: chiton

2010-05-16 Thread Sonny Heer
> site-packages. You might have to do something like this if you're on > Windows. > > set PYTHONPATH=C:\path\to\twisted > > More at... > > http://docs.python.org/tutorial/modules.html#the-module-search-path > > K > > > > On Fri, May 14, 2010 at 9:18 PM

Re: How to use binaryMemTable to insert large data?

2010-05-16 Thread Sonny Heer
What do you mean it didn't work? CassandraBulkLoader is a map/reduce program, so you'll need hadoop setup. On Sun, May 16, 2010 at 7:43 AM, Peng Guo wrote: > I try the contrib\bmt_example\CassandraBulkLoader.java Example, but it not > work. > > Can i try contrib\bmt_example\CassandraBulkLoader.j

Re: chiton

2010-05-14 Thread Sonny Heer
trac/wiki/Downloads > for the Windows/Ubuntu binaries. > Hope that helps, > Jon. > On 14 May 2010 21:04, Sonny Heer wrote: >> >> I'm having problems installing chiton... >> >> on ubuntu 9.10 I'm getting the following: >> Traceback (most recen

chiton

2010-05-14 Thread Sonny Heer
I'm having problems installing chiton... on ubuntu 9.10 I'm getting the following: Traceback (most recent call last): File "chiton/bin/chiton-client", line 2, in from twisted.internet import gtk2reactor ImportError: No module named twisted.internet does chiton work on windows?

Timeouts running batch_mutate

2010-05-14 Thread Sonny Heer
Hey, I'm running a map/reduce job, reading from HDFS directory, and reducing to Cassandra using the batch_mutate method. The reducer builds the list of rowmutations for a single row, and calls batch_mutate at the end. As I move to a larger dataset, i'm seeing the following exception: Caused by:

Deletion batch mutate

2010-05-06 Thread Sonny Heer
The Deletion Class only has a setSuper_column method. Does this work with regular columns as well? if not, how do you add a mutation for column delete?

Re: Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
C for ConcurrentMarkSweep: 12313 ms, 25638616 reclaimed leaving 1060647272 used; max is 1174208512 On Wed, Apr 21, 2010 at 3:16 PM, Jonathan Ellis wrote: > On Wed, Apr 21, 2010 at 5:05 PM, Sonny Heer wrote: >> What does OOM stand for? > > out of memory > >> for a given inse

Re: Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
what i mean by as data is processed is that the column size will grow in cassandra, but my client isn't ever writing large column size under a given row... Any idea whats going on here? On Wed, Apr 21, 2010 at 3:05 PM, Sonny Heer wrote: > What does OOM stand for? > > for a given i

Re: Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
wrote: > then that's not the problem. > > are you writing large rows that OOM during compaction? > > On Wed, Apr 21, 2010 at 4:34 PM, Sonny Heer wrote: >> They are showing up as completed?  Is this correct: >> >> >> Pool Name                    Act

Re: Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
ewhere that means you're just generating > insert ops faster than it can handle. > > On Wed, Apr 21, 2010 at 4:07 PM, Sonny Heer wrote: >> note: I'm using the Thrift API to insert.  The commitLog directory >> continues to grow.  The heap size continues to grow as wel

Re: Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
uot;insert threshold"? I moved to a more powerful node as well, it still ended up failing just after a longer period. On Wed, Apr 21, 2010 at 10:53 AM, Jonathan Ellis wrote: > http://wiki.apache.org/cassandra/FAQ#slows_down_after_lotso_inserts > > On Wed, Apr 21, 2010 at 12:

Import using cassandra 0.6.1

2010-04-21 Thread Sonny Heer
Currently running on a single node with intensive write operations. After running for a while... Client starts outputting: TimedOutException() at org.apache.cassandra.thrift.Cassandra$insert_result.read(Cassandra.java:12232) at org.apache.cassandra.thrift.Cassandra$Client.recv

Delete row

2010-04-20 Thread Sonny Heer
How do i delete a row using BMT method? Do I simply do a mutate with column delete flag set to true? Thanks.

Re: Map/Reduce Cassandra Output

2010-04-19 Thread Sonny Heer
antav/hector > > Thanks, > Stu > > -Original Message- > From: "Sonny Heer" > Sent: Monday, April 19, 2010 11:29am > To: cassandra-u...@incubator.apache.org > Subject: Map/Reduce Cassandra Output > > Different from the wordcount my input source is a directory

Map/Reduce Cassandra Output

2010-04-19 Thread Sonny Heer
Different from the wordcount my input source is a directory, and I have the a split class and record reader defined. Different from wordcount during reduce I need to insert into Cassandra. I notice for the wordcount input it retrieves a handle on a cassandra client like this: TSocket soc

Data model question - column names sort

2010-04-15 Thread Sonny Heer
Need a way to have two different types of indexes. Key: aTextKey ColumnName: aTextColumnName:55 Value: "" Key: aTextKey ColumnName: 55:aTextColumnName Value: "" All the valuable information is stored in the column name itself. Above two can be in different column families... Queries: Given a ke

Re: BMT flush on windows?

2010-04-15 Thread Sonny Heer
ption: java.util.concurrent.FutureTask" This may be known issue, just thought I'd pass it along. Not sure why using code from CassandraBulkLoader to send messages isn't working. I'm using Cassandra .6rc1. Thanks. On Thu, Apr 15, 2010 at 11:19 AM, Sonny Heer wrote: > Hm

Re: BMT flush on windows?

2010-04-15 Thread Sonny Heer
there is nothing to flush. > > On Thu, Apr 15, 2010 at 11:53 AM, Sonny Heer wrote: >> From the jconsole, I go under >> ColumnFamilyStores->CF1->Column1->Operations and clicked force >> flush(). >> >> I'm getting a "Operation return value"

Re: BMT flush on windows?

2010-04-15 Thread Sonny Heer
e options > > (a) connect with jconsole or another jmx client and invoke flush that way > (b) run org.apache.cassandra.tools.NodeCmd manually > (b) write a bat file for NodeCmd like the nodetool shell script in bin/ > > On Tue, Apr 13, 2010 at 5:08 PM, Sonny Heer wrote: >> Is there any way to run a keyspace flush on a windows box? >> >

Re: BMT flush on windows?

2010-04-13 Thread Sonny Heer
shell script in bin/ > > On Tue, Apr 13, 2010 at 5:08 PM, Sonny Heer wrote: >> Is there any way to run a keyspace flush on a windows box? >> >

BMT flush on windows?

2010-04-13 Thread Sonny Heer
Is there any way to run a keyspace flush on a windows box?

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
ssing something obvious here... On Thu, Apr 8, 2010 at 1:07 PM, Jonathan Ellis wrote: > those aren't shipped with Cassandra. > > On Thu, Apr 8, 2010 at 3:00 PM, Sonny Heer wrote: >> Missing the commons logging and commons httpclient jars.  Must be >> using the the wrong jd

Re: Cassandra .6 map/reduce

2010-04-08 Thread Sonny Heer
he script, Hadoop will use that configuration to > connect to your cluster. > > -Original Message----- > From: "Sonny Heer" > Sent: Thursday, April 8, 2010 3:06pm > To: cassandra-u...@incubator.apache.org > Subject: Cassandra .6 map/reduce > > Running the word_

Cassandra .6 map/reduce

2010-04-08 Thread Sonny Heer
Running the word_count example the hadoop job appears to be run internally. If I have a Cassandra cluster of 10 nodes, how does the Hadoop cluster get configured?

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
Missing the commons logging and commons httpclient jars. Must be using the the wrong jdk? On Thu, Apr 8, 2010 at 12:38 PM, Sonny Heer wrote: > Is there other documentation on how to setup all the pieces? > > Currently I'm simply trying to test the example word_count, but will &g

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
arted the default configuration. I checked out the .6rc1 project code from svn. 'cd' to contrib/word_count and trying to run the sample. Is this the correct way to run contrib stuff? where does the hadoop cluster come in? On Thu, Apr 8, 2010 at 12:18 PM, Sonny Heer wrote: > Okay

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
Okay I moved everything to the ubuntu box: ~/dev/cassandra-0.6.0-rc1/contrib/word_count$ bin/word_count_setup 10/04/08 11:15:10 INFO config.DatabaseDescriptor: Auto DiskAccessMode determined to be standard 10/04/08 11:15:10 WARN config.DatabaseDescriptor: KeysCachedFraction is deprecated: use Keys

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
e setting ThriftAddress to localhost on the Windows machine > precludes anything connecting. > > On Thu, Apr 8, 2010 at 11:58 AM, Sonny Heer wrote: >> Single node cluster (the windows box).  the Ubuntu box is only used to >> run the word count >> >> On Thu, Apr 8, 2010

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
27;s > your problem. > > On Thu, Apr 8, 2010 at 11:44 AM, Sonny Heer wrote: >> I have two boxes.  One is a windows box running Cassandra .6, and the >> other is an ubuntu box from which I'm trying to run the word count >> program as in the readme. >> >&g

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
n the > Windows machine? > > On Thu, Apr 8, 2010 at 11:44 AM, Sonny Heer wrote: >> I have two boxes.  One is a windows box running Cassandra .6, and the >> other is an ubuntu box from which I'm trying to run the word count >> program as in the readme. >> >&

Re: Iterate through entire data set

2010-04-08 Thread Sonny Heer
Wed, Apr 7, 2010 at 6:07 PM, Stu Hood wrote: > Please read the README in the contrib/word_count directory. > > -----Original Message- > From: "Sonny Heer" > Sent: Wednesday, April 7, 2010 6:33pm > To: user@cassandra.apache.org > Subject: Re: Iterate through entir

Re: Iterate through entire data set

2010-04-07 Thread Sonny Heer
Jon, I've got the word_count.jar and a Hadoop cluster. How do you usually run this sample? On Wed, Apr 7, 2010 at 3:04 PM, Jonathan Ellis wrote: > Yes > > On Wed, Apr 7, 2010 at 5:01 PM, Sonny Heer wrote: >> These examples work on Cassandra .06 and Hadoop .20.2? >> &

Re: Iterate through entire data set

2010-04-07 Thread Sonny Heer
These examples work on Cassandra .06 and Hadoop .20.2? On Wed, Apr 7, 2010 at 2:49 PM, Jonathan Ellis wrote: > Look at the READMEs for contrib/word_count and contrib/pig. > > On Wed, Apr 7, 2010 at 4:47 PM, Sonny Heer wrote: >> I need a way to process all of my data set. >>

Iterate through entire data set

2010-04-07 Thread Sonny Heer
I need a way to process all of my data set. A way to process every keyspace, CF, row, column, and perform some operation based on that mapped combination. The map bucket would collect down to column name. Is there a map/reduce program which shows how to go about doing this?

Re: Dividing the client load between machines in Cassandra

2010-03-17 Thread Sonny Heer
Opps. Yep, thanks! On Wed, Mar 17, 2010 at 1:47 PM, Jonathan Ellis wrote: > You didn't call tr.open() ? > > On Wed, Mar 17, 2010 at 3:45 PM, Sonny Heer wrote: >> I'm getting: >> org.apache.thrift.transport.TTransportException: Cannot write to nu

Re: Dividing the client load between machines in Cassandra

2010-03-17 Thread Sonny Heer
try { String jsonServerList = client.get_string_property("token map"); What am I doing wrong here? On Wed, Mar 17, 2010 at 11:33 AM, Sonny Heer wrote: > Cool thanks Todd.  I'd be interested at some point to see the updated > .6 version as well.  Thanks aga

Re: Dividing the client load between machines in Cassandra

2010-03-17 Thread Sonny Heer
} > //            } > //            catch ( JSONException e ) { > //                throw new ClusterRuntimeException( "Could not parse JSON > returned from Cassandra - don't know what to do?  A" ); > //            } > > > Sonny Heer wrote: >> &

Re: Dividing the client load between machines in Cassandra

2010-03-16 Thread Sonny Heer
et this to be an IP >> that is only for "node to node" communication.  The "ThriftAddress" param is >> the one i really want.  maybe this has been changed, "fixed" ;), haven't >> tested in a while. >> >> Jonathan Ellis wrote: >>&

Dividing the client load between machines in Cassandra

2010-03-16 Thread Sonny Heer
How can I accomplish this? The way I'm doing it now it is creating a TSocket connection using a static IP of one of the boxes on Cassandra: TTransport tr = new TSocket(host, port.intValue()); TProtocol proto = new TBinaryProtocol(tr); Cassandra.Client client = new Cassandra