Hi Ran, thanks for the compliment. It is true that we benefited enormously
from batch mutate. Without that the Mutator/Selector paradigm would not have
been possible in the same way. It will be interesting to see where Cassandra
takes us next. Best, Dominic
On 12 June 2010 20:05, Ran Tavory wrote
Also think this looks really promising.
The fact that there are so many API wrappers now (3?) doesn't reflect
well on the native API though :)
/me ducks and runs
On Mon, Jun 14, 2010 at 11:55, Dominic Williams
wrote:
> Hi Ran, thanks for the compliment. It is true that we benefited enormously
>
Can I expect batch_mutate to work in what I would think of as an atomic
operation?
That either all the mutations in the batch_mutate call are executed or none of
them are? Or can some of them fail while some of them succeeds?
>>
>> TBH while we are using super columns, the somehow feel wrong to me. I
>> would be happier if we could move what we do with super columns into
>> the row key space. But in our case that does not seem to be so easy.
>>
>>
>
> I'd be quite interested to learn what you are doing with super colu
no, it's not atomic. it just shortens the roundtrip of many update requests.
Some may fail and some may succeed
On Mon, Jun 14, 2010 at 2:40 PM, Per Olesen wrote:
> Can I expect batch_mutate to work in what I would think of as an atomic
> operation?
>
> That either all the mutations in the batch
Hi Riyad,
No problem. Because it is a new library, I cannot provide a large list of
production deployments. However, there are various reasons you should have
confidence in the library:-
1/ Firstly the background of Pelops is that it is being used as the basis of
a serious commercial project that
This question has been coming up quite regularly now. I've added an
entry to the FAQ. Please feel free to expand an clarify.
http://wiki.apache.org/cassandra/FAQ#batch_mutate_atomic
Gary.
On Mon, Jun 14, 2010 at 06:43, Ran Tavory wrote:
> no, it's not atomic. it just shortens the roundtrip of
Hi,
I have a question that relates to how to best model data. I have some pretty
simple tabular data, which I am to show to a large amount of users, and the
users need to be able to search some of the columns.
Given this tabular data:
Company| Amount|...many more columns here
--
I'll put something together and submit it. Thanks for the help.
Todd
-Original Message-
From: Gary Dusbabek [mailto:gdusba...@gmail.com]
Sent: Friday, June 11, 2010 4:49 AM
To: user@cassandra.apache.org
Subject: Re: Running Cassandra as a Windows Service
Sure. Please create a jira tic
That's the tradeoff we made to get basic functionality for a dozen or
so languages for free; it's impossible to be idiomatic with Thrift.
The glass-half-full view is, having lots of API wrappers shows that
building on Thrift is far easier than throwing bytes around at the
socket layer the way a tr
On Jun 13, 2010, at Sun Jun 13, 9:34 PM, Benjamin Black wrote:
> On Sun, Jun 13, 2010 at 5:58 PM, Matthew Conway wrote:
>> The ability to dynamically add new column families. Our app is currently
>> under heavy development, and we will be adding new column families at least
>> once a week aft
Great API that looks easy and intuitive to use. Regarding your connection pool
implementation, how does it handle failed/crashed nodes? Will the pool
auto-detect failed nodes via a "tester" thread or will a failed node, and hence
its pooled connection(s), be removed only when they are used? C
On Mon, Jun 14, 2010 at 7:27 AM, Matthew Conway wrote:
>
> On Jun 13, 2010, at Sun Jun 13, 9:34 PM, Benjamin Black wrote:
>
>> On Sun, Jun 13, 2010 at 5:58 PM, Matthew Conway wrote:
>>> The ability to dynamically add new column families. Our app is currently
>>> under heavy development, and we
On Jun 14, 2010, at Mon Jun 14, 10:45 AM, Jonathan Ellis wrote:
>>
>> I already have automation, whats missing are the details of the exact steps
>> I need to automate to accomplish the schema modification on a live cluster.
>> Even the FAQ just points to the feature in 0.7 trunk.
>
> Huh? h
Good to have new API wrappers.
I guess different APIs because people look Cassandra from different angle
with different backgrounds/ skills. At this stage, it's better that
different API find the good ideas from each other, and maybe one day, there
is one, which widely accepted. That's good for al
Hi, re: pools and detecting node failure...
Pooling is handled by ThriftPool. This class maintains a separate
NodeContext object for each known node. This in turn maintains a pool of
connections to its node.
Each NodeContext has a single "poolRefiller" object/thread, which runs
either when signal
Done, https://issues.apache.org/jira/browse/CASSANDRA-1188
It was find_all_by_service_id that was the culprit, and it resolves down to a
multiget_slice on a super column family. The super CF is acting as an index
back into a regular CF, thus I'm providing key, supercolumn name, and getting
bac
Hi,
thanks for your answer.
my comparators for this CF were:
CompareWith="LongType", CompareSubcolumnWith="UTF8Type" (which is not really
correct, as I use Longs for both, but I guess it should not cause such an
error).
I didn't test for empty start, and I already refactored the code not to do
On Mon, Jun 14, 2010 at 6:09 AM, Per Olesen wrote:
>
> So, in my use case, when searching on e.g. company, I can then access the
> "DashboardCompanyIndex" with a slice on its SC and then grab all the uuids
> from the columns, and after this, make a lookup in the Dashboard CF for each
> uuid fou
Hi everyone,
i am new to nosql databases and especially column-oriented Databases
like cassandra.
I am a student on information-systems and i evaluate a fitting no-sql
database for a web analytics system. Got the use-case of data like
webserver-logfile.
in an RDBMS it would be for every hit a row i
On Jun 14, 2010, at 6:29 PM, Benjamin Black wrote:
> On Mon, Jun 14, 2010 at 6:09 AM, Per Olesen wrote:
>>
>> So, in my use case, when searching on e.g. company, I can then access the
>> "DashboardCompanyIndex" with a slice on its SC and then grab all the uuids
>> from the columns, and after
Thank you for the very clear and detailed explanation of how the pool works.
I think I'll give Pelops a try.
Todd
From: Dominic Williams [mailto:thedwilli...@googlemail.com]
Sent: Monday, June 14, 2010 8:16 AM
To: user@cassandra.apache.org
Subject: Re: Pelops -
I checked out the source and noticed a few things:
1. You did not include an Ant build file. Not a big deal, but if you happen
to have one it would be nice to have.
2. It appears you built against Cassandra 0.6.0. Have you built and/or run
Pelops against 0.6.2 or trunk?
Todd
___
Hi Todd, we're on 6 or 6.1 but I would hope it should work with 6.2
We will also be moving to 6.2 very shortly because we need the column
timeout feature that a 6.2 patch has been submitted for (I haven't checked,
maybe that feature is actually in the 6.2 trunk now, would be great?)
In relation t
On Mon, Jun 14, 2010 at 2:28 PM, Dominic Williams
wrote:
> Hi Todd, we're on 6 or 6.1 but I would hope it should work with 6.2
> We will also be moving to 6.2 very shortly because we need the column
> timeout feature that a 6.2 patch has been submitted for (I haven't checked,
> maybe that feature
Hi,
I was updating to a newer 0.6.3 and happened to remember that I noticed
back in 0.6.2 there's this change in CHANGES.txt
* improve default JVM GC options (CASSANDRA-1014)
Looking at that ticket, I don't actually see the options listed or a
reason for why they changed. Also, I'm not certa
Hi,
I wrote 200k records to db with each record 5MB. Get this error when I uses
3 threads (each thread tries to read 200k record totally, 100 records a
time) to read data from db. The write is OK, the error comes from read.
Right now the Xmx of JVM is 1GB. I changed it to 2GB, still not working. I
Now I read 100 records each time, and the total time to read 200k records
(1M each) reduce to 10s. Looks good. But I am still curious how to handle
the case that users read one record each time,
On Fri, Jun 11, 2010 at 6:05 PM, Dop Sun wrote:
> And also, you are only select *1* key and *10* col
My guess: you are outrunning your disk I/O. Each of those 5MB rows
gets written to the commitlog, and the memtable is flushed when it
hits the configured limit, which you've probably left at 128MB. Every
25 rows or so you are getting memtable flushed to disk. Until these
things complete, they ar
"...or does it very greatly from installation to installation?"
Yes.
We're excited to announce Surge, the Scalability and Performance
Conference, to be held in Baltimore on Sept 30 and Oct 1, 2010. The
event focuses on case studies that demonstrate successes (and failures)
in Web applications and Internet architectures.
Our Keynote speakers include John Allspaw an
31 matches
Mail list logo