Hello,
I run Hadoop jobs which read data from Cassandra 1.2.8 and write results back
to another tables. One of my reduce tasks was killed 2 times by job tracker,
because it wasn't responding for more than 10 minutes, the 3rd attempt was
succesfull.
The error message for killed reduce tasks is
Hello,
We have a column family which stores incoming requests, and we would like to
perform some analytics on that data using Hadoop. The analytic results should
be available pretty soon, not realtime, but within an hour or so.
So we store the current hour number (calculated from timestamp) a
n how big your column is, you may also want to increase thrift
>>message length through setThriftMaxMessageLengthInMb().
>>
>>Hope that helps.
>>
>>On Tue, Sep 10, 2013 at 8:18 PM, Renat Gilfanov < gren...@mail.ru > wrote:
>>>Hi,
>>>
>>>
Hi,
We have Hadoop jobs that read data from our Cassandra column families and write
some data back to another column families.
The input column families are pretty simple CQL3 tables without wide rows.
In Hadoop jobs we set up corresponding WHERE clause in
ConfigHelper.setInputWhereClauses(...)
Hello,
Let's say we have a simple CQL3 table
CREATE TABLE example (
id UUID PRIMARY KEY,
timestamp TIMESTAMP,
data ASCII
);
And I need to mutate (for example encrypt) column values in the "data" column
for all rows.
What's the recommended approach to perform such migration progr
from seeds then?
Вторник, 3 сентября 2013, 14:08 -07:00 от Robert Coli :
>On Tue, Sep 3, 2013 at 2:01 PM, Renat Gilfanov < gren...@mail.ru > wrote:
>>
>>We have Cassandra cluster with 5 nodes hosted in the Amazon EC2, and I had
>>to restart two of them
Hello,
We have Cassandra cluster with 5 nodes hosted in the Amazon EC2, and I had to
restart two of them, so their IPs changed.
We use NetworkTopologyStrategy, so I simply updated IPs in the
cassandra-topology.properties file.
However, as I understood, old IPs remained somewhere in the system
Hello,
I'd like to ask what is the best options of separating commit log and data on
Amazon m1.xlarge instance, given 4x420 Gb attached storages and EBS volume ?
As far as I understand, the EBS is not the choice and it's recomended to use
attached storages instead.
Is it better to combine 4 ep
the cluster, then decommission
>the old ones, you'll be able to do it without downtime. It'll also have the
>effect of randomizing the tokens, I believe.
>
>On Sep 2, 2013, at 4:21 PM, Renat Gilfanov < gren...@mail.ru > wrote:
>
>> Hello,
>>
>> Currently
Hello,
Currently we have a Cassandra cluster in the Amazon EC2, and we are planning to
upgrade our deployment configuration to achieve better
performance and stability. However, a lot of open questions arise when planning
this migration. I'll be very thankfull if somebody could answer my
ques
10 matches
Mail list logo