>
> I have a patch for trunk which I just have to get time to test a bit before I
submit.
> It is for super columns and will use the super columns timestamp as the base
and only store variant encoded offsets in the underlying columns.
>
Could you please measure how much real benefit it brings
> in the family. There are millions of rows. Each operation consists of
> doing a batch_insert through pycassa, which increments ~17k keys. A
> majority of these keys are new in each batch.
>
> Each operation is taking up to 15 seconds. For our system this is a
> significant bottleneck.
>
Try t
This is golden! thanks a heap guys
On Mon, Sep 5, 2011 at 6:07 PM, Nick Bailey wrote:
> You can place each of the 4 new nodes exactly in the middle of 2 of
> the current nodes. This way each node will still be responsible for
> the same amount of data but your old nodes did not move.
>
> On Mon,
That issue says you can workaround the problem by turning off auto node
discovery... any instructions on how to do that? Is it done on the cluster
or just the sstableloader?
Thanks.
On Thu, Sep 1, 2011 at 5:34 PM, Jonathan Ellis wrote:
> Sounds like https://issues.apache.org/jira/browse/CASSAN
You can set up an instance of
org.scale7.cassandra.pelops.SimpleConnectionAuthenticator and pass it to
org.scale7.cassandra.pelops.IConnection.Config.
Cheers,
Dan
On Monday, 5 September 2011 at 4:24 PM, lacosa2...@libero.it wrote:
> Hi,
> I wanna know if exists and how to implement authentic
great thanks!
Evgeny.
You can place each of the 4 new nodes exactly in the middle of 2 of
the current nodes. This way each node will still be responsible for
the same amount of data but your old nodes did not move.
On Mon, Sep 5, 2011 at 2:56 PM, amulya rattan wrote:
> Ah, missed that. Thanks for the pointer.
> While
The first node can answer the question as long as you've requested
less rows than the first node has on it. Hence the "low cardinality"
point in what you quoted.
On Sat, Sep 3, 2011 at 5:00 AM, Kaj Magnus Lindberg
wrote:
> Hello Anyone
>
> I have a follow up question on a question from February
Hi Magnus,
I think the answer might be on
https://issues.apache.org/jira/browse/CASSANDRA-749. For example,
Jonathan writes:
> Is it worth creating a secondary index that only contains local data, versus
> a distributed secondary index (a normal ColumnFamily?)
I think my initial reasoning was
On Mon, 2011-09-05 at 21:52 +0200, Patrik Modesto wrote:
> I'm not sure about 0.8.x and 0.7.9 (to be released today with your
> patch) but 0.7.8 will fail even with RF>1 when there is Hadoop
> TaskTracer without local Cassandra. So increasing RF is not a
> solution.
This isn't true (or not the in
Ah, missed that. Thanks for the pointer.
While we are at it, the doc says that if I am doubling the strength of my
cluster, and I assign calculated tokens to the new nodes, i don't need to do
the nodetool move for old nodes. Won't I have to assign the old nodes with
their new respective tokens too?
On Mon, Sep 5, 2011 at 09:39, Mick Semb Wever wrote:
> I've entered a jira issue covering this request.
> https://issues.apache.org/jira/browse/CASSANDRA-3136
>
> Would you mind attaching your patch to the issue.
> (No review of it will happen anywhere else.)
I see Jonathan didn't change his mind
https://issues.apache.org/jira/browse/CASSANDRA-3082
On Mon, Sep 5, 2011 at 10:04 AM, Evgeniy Ryabitskiy
wrote:
> Hi,
>
> I'am trying to store record with EACH_QUORUM consistency and RF=3. While
> same thing with RF=2 is working.
> Could some one tell me why EACH_QUORUM is working with RF=2 but n
Hi,
I am facing some problem using Thrift 7.
I downloaded the tar file.I downloaded the windows exe too for . Created a
thrift jar from the lib java path and then generated the java class from
tutorial.thrift file.Now while I use run the ant file inside example I got
some error that some of the
I'm interested in isolating the replication logic, so that cassandra (or any
NoSQL software, for that matter) is composed of
the replication module and DB engine, and possibly key mapping module.
this way we could swap out Cassandra replication (multi-master, async) with
, for example, ZAB protoco
On Mon, 2011-09-05 at 19:02 +0200, Mick Semb Wever wrote:
>
> ConfigHelper.setInputRange(
> jobConf,
>
> partitioner.getTokenFactory().toString(partitioner.getToken(myKey)),
>
> partitioner.getTokenFactory().toString(partitioner.getToken(my
Hey guys,
We are designing our data model for our app and this question came up.
Lets say that I have a large number of rows(say 1M). And just one column
family.
Each row contains either columns (A, B, C) or (X, Y, Z). I want to run a
get_range_slices query to fetch columns (A, B, C).
Does cassandr
One more thing, Cassandra version is 0.8.4.
And if I try same thing from Pelops(thrift), I get UnavailableException.
On Mon, 2011-09-05 at 18:18 +0300, Vitaly Vengrov wrote:
> See these rows in the ColumnFamilyInputFormat.getSplits method :
>
> assert jobKeyRange.start_key == null : "only start_token supported";
>
> assert jobKeyRange.end_key == null : "only end_token supported";
>
> So, the
Hi guys.
See these rows in the ColumnFamilyInputFormat.getSplits method :
assert jobKeyRange.start_key == null : "only start_token
supported";
assert jobKeyRange.end_key == null : "only end_token
supported";
So, the question is why start_key and end_key aren't sup
Hi,
I'am trying to store record with EACH_QUORUM consistency and RF=3. While
same thing with RF=2 is working.
Could some one tell me why EACH_QUORUM is working with RF=2 but not with RF
>=3
I have 7 nodes cluster. All nodes are UP.
Here is simple CLI script:
create keyspace kspace3
with placeme
Yes, my cluster is working.
I didn't realize it at the time, but the StorageService link I listed is
already in 0.8.4, so yes the only file I had to patch was VersionedValue.
Not sure what was going on with the pig jars, but after more configuration
changes than I can count, I'm pretty sure remov
Hi. We got the same problem here. Even the wordcount map/reduce example in
the source tar works fine with one node, but fails with the same exception
on a two node cluster. CASSANDRA-3044 mentioned that a temporary work around
is to disable node auto discovery. Can anyone tell me how to do that in
Have you done step 6 of the 'To add nodes to a Cassandra cluster' of
http://www.datastax.com/docs/0.8/operations/clustering#adding-capacity,
aka, run nodetool cleanup on the previously existing nodes ?
--
Sylvain
On Sun, Sep 4, 2011 at 11:58 AM, amulya rattan wrote:
> Hi there,
> I had a 3 nodes
On Fri, 2011-09-02 at 09:28 +0200, Patrik Modesto wrote:
> We use Cassandra as a storage for web-pages, we store the HTML, all
> URLs that has the same HTML data and some computed data. We run Hadoop
> MR jobs to compute lexical and thematical data for each page and for
> exporting the data to a bi
25 matches
Mail list logo