Hope you find following useful. It uses raw thirft. In case you find
difficulty in build and/or running the code, please reply back.
private Cassandra.Client createClient(String host, int port) {
TTransport framedTransport = new TFramedTransport(new TSocket(host,
port));
TProtocol framedPr
> Based on what you're saying, and being I'm using session (cookie) based load
> balancing it would be true that data is rarely read or written (per user) on
> a different server, that could be why data isn't replicating.
You've probably discovered this already but just in case, and for
others f
I am running a get_range_slices on one of my larger CFs. I am then
running a 'get' call on each of those keys. I have run into 50 or so
keys that were returned in the range, but get a NotFound when called
against 'get'.
I repeated the range call to ensure they weren't simply recently
modified/dele
Nvm. Found the answer in the FAQ :P It is normal.
Thx,
Jason
On Fri, Mar 25, 2011 at 1:24 AM, Jason Harvey wrote:
> I am running a get_range_slices on one of my larger CFs. I am then
> running a 'get' call on each of those keys. I have run into 50 or so
> keys that were returned in the range, bu
Hi All,
I am evaluating Cassandra as a data store for my application.
This is what the format of the data I want to store looks like:
{
timestampuuid: unique time stamp
finite_set_of_values_attribute1: this is a value from a finite set ,
lets say {a,b,c,d}
finite_set_of_values_attribut
2011/3/25 Jonathan Ellis :
> You can't have them both in the same Mutation struct, but you can have
> them in the map>> that is the "batch."
Actually you can't, because we created a different thrift structure
(CounteMutation) to hold
counter mutations. Looking back, we really just did that to have
On Fri, Mar 25, 2011 at 1:49 AM, Jared Laprise wrote:
> Hello all, I’m running 2 Cassandra 6.5 nodes and I brought down the
> secondary node and restarted the primary node. After Cassandra came back up
> all data has been reverted to several months ago.
Out of curiosity, when you said 'brought do
Hi everyone,
I'm running 10 cassandra nodes in total and updated 5 nodes
from 0.7.0 to 0.7.2 one by one like below.
# sudo /etc/init.d/cassandra stop
# sudo yum update cassandra-0.7.2-4
# sudo /etc/init.d/cassandra start
After updating, those 5 updated nodes keep outputting error messages to
/v
That kind of aggregation is certainly possible today, programmatically...
but if you want to do it in cassandra only, you are out of luck, today.
But it sounds like the project DataStax just announced might help quite a
bit with a use case like that.
David
Sent from my Droid
On Mar 25, 2011 3:58
May I point to complex event processing, see http://esper.codehaus.org. Esper
provides a SQL-like language for performing the queries, the engine internally
maintains the indexes for answering queries fast. EsperHA runs on Cassandra.
Best regards,
Tom
project lead Esper
___
Hi,
from the node log that has inconsistent view -there is
INFO [ScheduledTasks:1] 2011-03-25 10:27:31,089 Gossiper.java (line 224)
InetAddress /XXX is now dead.
But from that node i can telnet/ping XXX on 9160,7000
Thats strange.
2011/3/23 Dave Viner
> I saw this once when my servers r
Hi Narendra,
Thanks for your email Which thrift version are you using.I
am using thrift-0.0.5 in which batch_mutate takes
map> as argument
in ur sample code ur map is of type map>
Please Suggest
Thanks
Anurag
On Fri, Mar 25, 2011 at 12:10 AM, Narendra Sharma wrote:
> Hope you
from NEWS.txt:
Upgrading
-
- 0.7.1 and 0.7.2 shipped with a bug that caused incorrect row-level
bloom filters to be generated when compacting sstables generated
with earlier versions. This would manifest in IOExceptions during
column name-based queries. 0.7.3 provid
Hi All,
I am using cassandra 0.7.3 and thrift 0.0.5 when I am trying to
write to cassandra using batch_mutate I get the following error:
InvalidRequestException(why:Key may not be empty)
at
org.apache.cassandra.thrift.Cassandra$batch_mutate_result.read(Cassandra.java:16007)
Hello Saurabh,
I have a similar situation, with a more complex data model, and I do an
equivalent of map-reduce "by hand". The redeeming value is that you have
complete freedom in how you hash, and you design the way you store indexes
and similar structures. If there is a pattern in data store, yo
Thanks everyone this gives me a good head start.
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Central-monitoring-of-Cassandra-cluster-tp6205275p6208331.html
Sent from the cassandra-u...@incubator.apache.org mailing list archive at
Nabble.com.
ok, I will test again and let you know.
SD - to secure data stream between EC2 regions, can we not just setup
a VPN in EC2 with this patch ?
On Wed, Mar 23, 2011 at 8:50 PM, Milind Parikh wrote:
> My nodetool repair does not hang. That's why I'm curious.
>
> /***
> sent from
Thanks, Jonathan!
Part of what we're trying to accomplish is a data cleanup. One of our
nodes seems to have some lingering data from an old column family that
we no longer have defined (we're running v0.60) so that node has a few
GB of data that never gets replicated. We're hoping that by brin
On Fri, Mar 25, 2011 at 11:59 AM, ian douglas wrote:
> Part of what we're trying to accomplish is a data cleanup. One of our nodes
> seems to have some lingering data from an old column family that we no
> longer have defined (we're running v0.60)
I don't know if you could hear that from where yo
could you paste the code that does that?
On Fri, Mar 25, 2011 at 6:47 AM, Anurag Gujral wrote:
> Hi All,
> I am using cassandra 0.7.3 and thrift 0.0.5 when I am trying to
> write to cassandra using batch_mutate I get the following error:
> InvalidRequestException(why:Key may not be empt
thank you both. That makes Hector design much easier as well.
On Fri, Mar 25, 2011 at 2:50 AM, Sylvain Lebresne wrote:
> 2011/3/25 Jonathan Ellis :
> > You can't have them both in the same Mutation struct, but you can have
> > them in the map>> that is the "batch."
>
> Actually you can't, because
Glad it worked
2011/3/25
> very cool. thanks for the info. this is exactly what we need.
>
>
> On Mar 25, 2011 8:22am, Patricio Echagüe wrote:
> >
> > It's a cassandra consistency level
> > On Mar 24, 2011 11:44 PM, jonathan.co...@gmail.com> wrote:> Patricio -
> > >
> > > I haven't heard of loc
2011/3/25 Eric Gilmore
> Also:
> http://www.datastax.com/docs/0.7/operations/clustering#adding-capacity
>
> Can do that about i represent, but i afraid that when i begin balance
cluster with new node this will be a big stress for it. Mey be exists some
strategies how to do that?
One thing I wonder though - if your columns are the thing that are increasing
your heap size and eating up a lot of memory, and you're reading the data
structure out as a bag of columns, why isn't pig spilling to disk instead of
growing in memory. The pig model is that you can have huge bags th
On 03/25/2011 10:12 AM, Jonathan Ellis wrote:
On Fri, Mar 25, 2011 at 11:59 AM, ian douglas wrote:
(we're running v0.60)
I don't know if you could hear that from where you are, but our whole
office just yelled, "WTF!" :)
Ah, that's what that noise was... And yeah, we know we're way behind.
Thanks for all the responses.
My leading questions then are ->
- Should I go with the OrderPreservingPartitioner based on timestamps so I
can do time range queries - is this recommended ? any special cases
regarding load balancing I need to keep in mind ? I have read buzz over
blogs/forums on how
I don't think it's Pig running out of memory, but rather Cassandra itself (the
data doesn't even make it to Pig). get_range_slices() is called with a row
batch size of 4096, the default, and it's fetching all of the columns in each
row. If I have 10K columns in each row, that's a huge request, a
No, what initially started it all was that I needed to increase my EC2 server
instance size. So I removed said server from the load balancer, stopped
Cassandra, and then shutdown the server in order to change the instance type. I
assumed the other node had all the data and everything should keep
> Although after all the help from the Cassandra community I have a much better
> understanding of why and how my situation happened, there was still one
> strange side effect I noticed. For context, I store user accounts and other
> account information in Cassandra. When the second node was off
Hey Jonathan,
Thanks for the response. I applied the patch to 0.7.4 and things have
started working again nicely.
Looks like this fix is going in 0.7.5. Any idea when 0.7.5 will be released?
Thanks again!
On Wed, Mar 23, 2011 at 9:56 PM, Jonathan Ellis wrote:
> This looks like a bug
> (https://i
Ruslan, I'm not sure exactly what risks you are referring to -- can you be
more specific?
Do the CPU-intensive operations one at a time, including doing the cleanup
when it will not interfere with other operations, and I think you should be
fine, from my understanding.
1. Start the new nodes
The Ubuntu Software Update seems to complain --
Failed to fetch
http://www.apache.org/dist/cassandra/debian/dists/unstable/main/binary-amd64/Packages.gz
403 Forbidden [IP: 140.211.11.131 80]
Failed to fetch
http://www.apache.org/dist/cassandra/debian/dists/unstable/main/source/Sources.gz
403 Forb
Does anyone know how cassandra chooses the nodes for its other replicant copies?
The first node gets the first copy because its token is assigned for that key.
But what about the other copies of the data?
Do the replicant nodes stay the same based on the token range? Or are the
other copies
2011/3/25 Eric Gilmore
> Ruslan, I'm not sure exactly what risks you are referring to -- can you be
> more specific?
>
> Do the CPU-intensive operations one at a time, including doing the cleanup
> when it will not interfere with other operations, and I think you should be
> fine, from my underst
Cassandra issue tracker have ticket for this (and in this list link on this
ticket was posted, but i forgot where)
2011/3/25 Shashank Tiwari
> The Ubuntu Software Update seems to complain --
> Failed to fetch
> http://www.apache.org/dist/cassandra/debian/dists/unstable/main/binary-amd64/Packages
> Does anyone know how cassandra chooses the nodes for its other replicant
> copies?
This keeps coming up so I added a FAQ entry:
http://wiki.apache.org/cassandra/FAQ#replicaplacement
I don't quite like the phrasing but I couldn't come up anything that
was sufficiently clear and complete rig
Just to be clear, it's also the case that if I have a Hadoop TaskTracker
running on each node that Cassandra is running on, a map/reduce job will
automatically handle data locality, right? I.e. each mapper will only read
splits which live on the same box.
-Jeffrey
-Original Message-
Fr
Yes. For your other question, I'm not sure but it makes sense that the
Cassandra memory usage would be separate from the pig memory usage - so pig my
be doing the spill to disk.
On Mar 25, 2011, at 6:21 PM, Jeffrey Wang wrote:
> Just to be clear, it's also the case that if I have a Hadoop Task
With RF=2 & CL=ONE, take care on that you still have chance to read old data
which is not replicated yet.
Maki
From iPhone
On 2011/03/26, at 5:10, Jared Laprise wrote:
> No, what initially started it all was that I needed to increase my EC2 server
> instance size. So I removed said server f
On Fri, Mar 25, 2011 at 1:41 PM, Jeffrey Wang wrote:
> I don't think it's Pig running out of memory, but rather Cassandra itself
> (the data doesn't even make it to Pig). get_range_slices() is called with a
> row batch size of 4096, the default, and it's fetching all of the columns in
> each ro
I had another question that ties in with my requirement.
How efficient is it to move data from one column family to another column
family ?
Basically, what I want to do is keep track of how "old" a certain data point
is. I have one column family that maintain those data points, and I want to
be a
People: Crapssandra.
I'm starting a Cassandra project and starting to learn about this beautiful
Cassandra, so I thougth that it would be nice to have a db gui tool under my
current OS.
It doesn't do anything other than showing some info about the server or the
selected keyspace... but I hope it'
42 matches
Mail list logo