Hi Folks,
I'm trying to wrap my brain around how one would configure a Cassandra
cluster to support consistent reads in various cluster failure states.
(IE: Reads must see the most recent write.)
Gracefully handling a single node failure seems easy enough. Use
QUORUM (RF/2+1) for reads and writ
Hi,
When I run nodetool compactionstats
I see the number of pending tasks keep going up steadily.
I tried to increase the compactionthroughput, by using
nodetool setcompactionthroughput
I even tried the extreme to set it to 0 to disable the throttling.
I checked iostats and we have SSD for
The fact that it's still exactly 521 seconds is very suspicious. I can't
debug your script over the mailing list, but do some sanity checks to make
sure there's not a bottleneck somewhere you don't expect.
On Fri, Jan 18, 2013 at 12:44 PM, Pradeep Kumar Mantha wrote:
> Hi,
>
> Thanks Tyler.
>
Hi,
Thanks Tyler.
Below is the *global* connection pool I am trying to use, where the
server_list contains all the ips of 12 DataNodes I am using and
pool_size is the number of threads and I just set to timeout to 60 to
avoid connection retry errors.
pool = pycassa.ConnectionPool('Blast',
serve
Network topology comes in picture when you have huge data to travel... you
need to consider the network aspects too...
∞
Shashwat Shriparv
On Fri, Jan 18, 2013 at 11:48 PM, Francisco Sobral wrote:
> Dear friends,
>
> We have only one datacenter with 4 nodes and no plans to have more
> datace
Dear friends,
We have only one datacenter with 4 nodes and no plans to have more datacenters.
With respect to the replication strategy, in this case, will SimpleStrategy be
more efficient than NetworkTopologyStrategy, since the the latter performs an
additional search for different racks?
Best
> Silly question -- but does hive/pig hadoop etc work with cassandra
> 1.1.8? Or only with 1.2?
all versions.
> We are using astyanax library, which seems
> to fail horribly on 1.2,
How does it fail ?
If you think you have a bug post it at https://github.com/Netflix/astyanax
Cheers
---
Hi Naveen,
You can start with http://wiki.apache.org/cassandra/HadoopSupport but there's
also a commercial product that you can use, DataStax Enterprise:
http://www.datastax.com/docs/datastax_enterprise2.2/solutions/hadoop_index
which makes things more streamlined, but it's a commercial product
You just need to increase the ConnectionPool size to handle the number of
threads you have using it concurrently. Set the pool_size kwarg to at
least the number of threads you're using.
On Thu, Jan 17, 2013 at 6:46 PM, Pradeep Kumar Mantha
wrote:
> Thanks Tyler.
>
> I just moved the pool and cf
It looks like snappy was the problem. Solved by setting
org.xerial.snappy.tempdir as detailed in
http://www.datastax.com/docs/1.2/troubleshooting/index#snappy
Thanks,
Traian.
On Fri, Jan 18, 2013 at 3:28 PM, Gabriel Ciuloaica wrote:
> also check firewalls(iptables, selinux) on the boxes ...
>
>
also check firewalls(iptables, selinux) on the boxes ...
Chers,
Gabi
On Jan 18, 2013, at 3:26 PM, Gabriel Ciuloaica wrote:
> you need snappy jna on that machine installed... at least this could be seen
> from the logs posted ...
>
> Gabi
>
> On Jan 18, 2013, at 12:19 PM, Traian Fratean
> wr
you need snappy jna on that machine installed... at least this could be seen
from the logs posted ...
Gabi
On Jan 18, 2013, at 12:19 PM, Traian Fratean
wrote:
> Hi,
>
> I have a cluster of Cassandra 1.2.0 running on latest stable Scientific Linux.
> Nodes start but when I check with nodetool
Hi,
I have a cluster of Cassandra 1.2.0 running on latest stable Scientific
Linux.
Nodes start but when I check with nodetool I see only local machine in the
ring.
$ /usr/bin/nodetool -host 10.60.15.63 -p 7199 ring
> Note: Ownership information does not include topology; for complete
> informatio
The Cassandra team is pleased to announce the release of Apache Cassandra
version 1.1.9.
Cassandra is a highly scalable second-generation distributed database,
bringing together Dynamo's fully distributed design and Bigtable's
ColumnFamily-based data model. You can read more here:
http://cassand
Priam is good for backups but it is another complex (but very good) part to a
software stack.
A simple solution is to do regular snapshots (via cron)
Compress them and put them into s3
On the s3 you can simply choose how many days the files are kept.
This can be done with a couple of lines of she
15 matches
Mail list logo