You can use expiring columns to say "only the last N seconds' of
data." But not counts per se for the reasons Aaron gave.
On Mon, Jul 18, 2011 at 10:22 AM, JKnight JKnight wrote:
> Dear all,
> I want to keep only 100 column of a key: when I add a column for a key, if
> the number column of key i
FYI - This manual reordering of the DCs and RACs might make it easier to see
how the tokens are arranged. Pretty sure that the token ranges are picked
correctly. Ignore the Owns column, b/c it is not multi-datacenter aware (so
it thinks all of the nodes are in one ring as opposed to two (DC1 & DC2)
We're running brisk v1 beta2 on 12 nodes - 8 cassandra in DC1 and 4 brisk in DC
2 in EC2. Wrote a few TBs of data to the cluster, and unfortunately the load is
very unbalanced. Every key is the same size and we are using RandomPartitioner.
There are two replicas of data in DC1 and one replica in
The Datastax post on counters says:
"Prior to 0.8, Cassandra had no simple and efficient way to count. By
“counting,” we mean here to provide an atomic increment operation in a
single column value, as opposed to counting the number of columns in a row,
or rows in a column family, both of which were
There is no support for a feature like that, and i doubt it would ever be
supported. For one there there are no locks during a write, so it's not
possible to definitively say there are 100 columns at a particular instance of
time.
You would need to read all columns and delete the ones you no l
Why are you installing thrift ?
The cassandra binary packages contain all the dependancies.
Cheers
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 19 Jul 2011, at 07:51, Sal Lopez wrote:
> Does anyone have documentation/tips for instal
> Of course. I wasn't thinking clearly.
>
> So, back to a previous point you brought up, I will have heavy reads and
> even heavier writes. How would you rate the benefits of flashcache in
> such a scenario? Is it still an overall performance boost worth the
> expense?
We have also heavy r
Your comparator is looking for the empty string as the first component.
For what you are trying to do, you probably want the TimeUUID to be
first so you can specify the range for the first positions, then ""
and "|" (or similarly brute-ish ascii search hack which would include
all the values of th
Indexes are not supported on sub columns.
Also, you definition seems to mix standard and sub columns together in the CF.
For a super CF all top level columns contain sub columns.
Cheers
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 14
org.apache.cassandra.db.marshal.DoubleType should work, then. the
cast shouldn't be necessary if the doubletype is actually in the
schema either.
On Mon, Jul 18, 2011 at 4:34 PM, Sameer Farooqui
wrote:
> Yup, Jonathan, I was just temporarily returning the value as LongType b/c I
> couldn't find
Yup, Jonathan, I was just temporarily returning the value as LongType b/c I
couldn't find a DoubleType. Since LongType returned something, I know that
the data is there.
Thanks for that link, Jeremy. I'm guessing DoubleType hasn't found its way
into the CLI yet, b/c I tried this and got an error:
"get X as LongType" will give you the bytes in that column value
interpreted as a long.
So it did exactly what you asked it to do :)
On Mon, Jul 18, 2011 at 4:16 PM, Sameer Farooqui
wrote:
> I wrote some data to a standard column family (using Hector) with ASCII
> row-key, Long column name and D
For the "Too many open files" error see:
http://www.datastax.com/docs/0.8/troubleshooting/index#java-reports-an-error-saying-there-are-too-many-open-files
Restart the node and see if the node is able to complete the pending repair
this time. Your node may have just been stuck on this error that ca
I know additional types have been added as of 0.8.1:
https://issues.apache.org/jira/browse/CASSANDRA-2530
However, I'm not sure how those have propagated up to validators, the CLI, and
hector though.
On Jul 18, 2011, at 4:16 PM, Sameer Farooqui wrote:
> I wrote some data to a standard column fa
I wrote some data to a standard column family (using Hector) with ASCII
row-key, Long column name and Double column value.
The CLI doesn't seem to have a double data type to interpret the column
value for display.
[default@DemoKS] get DemoCF[ascii('ABC_1020')][long(2009010100)]
as Lon
I'm running into a quirky issue with Brisk 1.0 Beta 2 (w/ Cassandra 0.8.1).
I think the last node in our cluster is having problems (10.201.x.x).
OpsCenter and nodetool ring (run from that node) show the node as down, but
the rest of the cluster sees it as up.
If I run nodetool ring from one of t
Does anyone have documentation/tips for installing thrift on a server that
does not have access to the internet? See error below:
Buildfile: build.xml
setup.init:
[mkdir] Created dir: /tmp/thrift-0.6.1/lib/java/build
[mkdir] Created dir: /tmp/thrift-0.6.1/lib/java/build/lib
[mkdir] Cr
On 7/18/2011 1:20 PM, Héctor Izquierdo Seliva wrote:
If using the version that has both rt and wt caches, is it just the wt
cache that's polluted for compactions/flushes? If not, why does the rt
cache also get polluted?
As I said, all reads go through flashcache, so if you read three 10 GB
sst
>
> If using the version that has both rt and wt caches, is it just the wt
> cache that's polluted for compactions/flushes? If not, why does the rt
> cache also get polluted?
>
As I said, all reads go through flashcache, so if you read three 10 GB
sstables for a compaction you will get those
It's okay, it seems that you need to enclose the comparator with single
quotes.
On another note, can ranges be set without based on just one components of a
composite column name?
I have a the Composite below in the previous email and want to return all my
transaction columns for a particular
> Thanks! Then does it mean that before compaction if read call comes for that
> key sort is done at the read time since column b, c and a are in different
> ssTables.
Essentially yes; a merge-sort happens (since they are sorted locally
in each sstable).
--
/ Peter Schuller (@scode on twitter)
On 7/18/2011 12:08 PM, Héctor Izquierdo Seliva wrote:
Interesting. So, there is no segregation between read and write cache
space? A compaction or flush can evict blocks in the read cache if it
needs the space for write buffering?
There are two versions, the -wt (write through) that will cache
Thanks! Then does it mean that before compaction if read call comes for that
key sort is done at the read time since column b, c and a are in different
ssTables.
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/How-are-column-sort-handled-tp6595415
> Interesting. So, there is no segregation between read and write cache
> space? A compaction or flush can evict blocks in the read cache if it
> needs the space for write buffering?
There are two versions, the -wt (write through) that will cache also
what is written, and the normal version t
Aaron,
That additional 15 bytes of overhead was the missing puzzle piece.
We had RF = 3.
So, now my calculations show that our CF should have a total of about 3.1 TB
of data and the actual figure is 3.3 TB (which might just be some stale
tombstones).
Thanks for the clarification about what else
On 7/18/2011 4:14 AM, Héctor Izquierdo Seliva wrote:
Hector, some before/after numbers would be great if you can find them.
Thanks!
I'll try and get some for you :)
What happens when your cache gets trashed? Do compactions and flushes
go slower?
If you use flashcache-wt flushed and compact
Is this possible? I'm trying to create a Column Family with a CompositeType
comparator, but I keep getting an error:
create column family Transactions with comparator = CompositeType(UTF8Type,
UUIDType) and keys_cached=1 and rows_cached=1000;
Syntax error at position 84: missing EOF at '('
Sa
> Trying to understand the overhead when multiple columns are spread accross
> ssTables. For eg: Key K1 column b and c are in ssTable 1 and column a in
> ssTable 2. As I understand columns in a given row are sorted at the time
> it's stored. So does it mean that when "a" goes to ssTable 2 it also f
Trying to understand the overhead when multiple columns are spread accross
ssTables. For eg: Key K1 column b and c are in ssTable 1 and column a in
ssTable 2. As I understand columns in a given row are sorted at the time
it's stored. So does it mean that when "a" goes to ssTable 2 it also fetches
c
No, the point is that it's the same memory that would be cached anyway
if you were using non-mmap'd I/O.
This will be more obvious once you have say 10x more data (and SHR)
than you have ram (and RES).
On Mon, Jul 18, 2011 at 10:27 AM, JKnight JKnight wrote:
> Thank for your response.
> Can we r
Thank for your response.
Can we reduce that value? Memory is used just 600M but the process occupy
3.2G. Too waste.
On Mon, Jul 18, 2011 at 6:53 PM, Jonathan Ellis wrote:
> That means that the mmaped files are indeed resident at the moment.
>
> On Mon, Jul 18, 2011 at 1:51 AM, JKnight JKnight
Dear all,
I want to keep only 100 column of a key: when I add a column for a key, if
the number column of key is 100, another column (by order) will be deleted.
Does Cassandra have setting for that?
--
Best regards,
JKnight
It would be nice if this were fixed before I move up to 0.8...
On Mon, Jul 18, 2011 at 3:19 PM, Boris Yen wrote:
> If it would not cause the dev team to much trouble, I think the cassandra
> should maintain the backward compatability regarding the generation of the
> default index_name, otherwis
If it would not cause the dev team to much trouble, I think the cassandra
should maintain the backward compatability regarding the generation of the
default index_name, otherwise when people start dropping columns indices,
the result might not be what they want.
On Mon, Jul 18, 2011 at 7:59 PM, Jo
On Mon, Jul 18, 2011 at 12:20 AM, Boris Yen wrote:
> Will this have any side effect when doing a get_indexed_slices
No
> or when a
> user wants to drop an index by any means?
Sort of; one of the indexes with the name will be dropped, but not all.
--
Jonathan Ellis
Project Chair, Apache Cassan
That means that the mmaped files are indeed resident at the moment.
On Mon, Jul 18, 2011 at 1:51 AM, JKnight JKnight wrote:
> Thank for your response.
> Do you talk about virtual memory (column VIRT show in top command)?
> But I mention about column RES. In my case, VIRT is 61.8G, RES is 3.2G and
Ah, that's it. I'm on 0.7
On Mon, Jul 18, 2011 at 1:27 PM, Boris Yen wrote:
> which version of cassandra do you use? What I mentioned here only happens
> on 0.8.1.
>
>
> On Mon, Jul 18, 2011 at 4:44 PM, David Boxenhorn wrote:
>
>> I have lots of indexes on columns with the same name. Why don't I
which version of cassandra do you use? What I mentioned here only happens on
0.8.1.
On Mon, Jul 18, 2011 at 4:44 PM, David Boxenhorn wrote:
> I have lots of indexes on columns with the same name. Why don't I have this
> problem?
>
> For example:
>
> Keyspace: City:
> Replication Strategy: org
>
> Hector, some before/after numbers would be great if you can find them.
> Thanks!
>
I'll try and get some for you :)
> What happens when your cache gets trashed? Do compactions and flushes
> go slower?
>
If you use flashcache-wt flushed and compacted sstables will go to the
cache.
A
http://permalink.gmane.org/gmane.comp.db.cassandra.user/14225
but given
https://issues.apache.org/jira/browse/CASSANDRA-2868
and me thinking 2 secs longer I guess it was the leaked native memory from gc
inspector that has been swapped out.
(I didn't believe that mlockall is broken but at that
I have lots of indexes on columns with the same name. Why don't I have this
problem?
For example:
Keyspace: City:
Replication Strategy: org.apache.cassandra.locator.SimpleStrategy
Replication Factor: 3
Column Families:
ColumnFamily: AttractionCheckins
Columns sorted by: org.apac
I haven't used CQL functionality much, but thirft client
I think what I encounter is exactly this problem!
>
If you want to query over key, you can index keys to other CF, get the
column names (that is key of other CF ). and then query actual CF with keys.
switch away from the random partitioner.
42 matches
Mail list logo