Re: [RELEASE] Apache Cassandra 1.1.11

2013-04-24 Thread Patrik Modesto
It works now, thanks! P. On Wed, Apr 24, 2013 at 2:53 PM, Sylvain Lebresne wrote: > Sorry, seems I screwed up somehow. > That should be fixed now however.

Re: [RELEASE] Apache Cassandra 1.1.11

2013-04-24 Thread Patrik Modesto
Hi! I can't find the 1.1.11 package for the Debian at http://www.apache.org/dist/cassandra/debian/ The .deb package is there but Packages files still contain just version 1.1.10. Regards, Patrik

Re: random partitioner and key scan

2012-07-20 Thread Patrik Modesto
On Fri, Jul 20, 2012 at 11:17 AM, aaron morton wrote: > Ordering the rows by row key locally would mean that every row on the node > would have to be scanned to find the ones whose token was in the required > token range. I don't know much about Cassandra internals, but from a user point of view,

Re: random partitioner and key scan

2012-07-19 Thread Patrik Modesto
ote: > The probem could be while redistributing the tokens. In that case the > hashes has to be recalculated on each fo the candidate node. > > -Thanks, > Prasenjit > > On Thu, Jul 19, 2012 at 12:19 PM, Patrik Modesto > wrote: >> Hi, >> >> I know that Ran

random partitioner and key scan

2012-07-18 Thread Patrik Modesto
Hi, I know that RandomPartitioner does MD5 of a key and the MD5 is then used for key distribution AND key ordering. I was just wondering if it's possible to have RandomPartitioner just for key distribution and OrderedPartitioner just for per-node key ordering. That would solve the often requested

Re: Removing a counter columns using Thrift interface

2012-06-26 Thread Patrik Modesto
On Mon, Jun 25, 2012 at 9:28 AM, Sylvain Lebresne wrote: > On Mon, Jun 25, 2012 at 9:06 AM, Patrik Modesto > wrote: >> I'm used to use Mutation for everything, so the first thing I tried >> was Deletion on Counter column. Well, nothing happened. No error and >> the C

Removing a counter columns using Thrift interface

2012-06-25 Thread Patrik Modesto
Hi, I was playing with Counters in Cassandra 1.0.10 and I see a behaviour that I didn't expect. It's about removing a Counter column. I'm used to use Mutation for everything, so the first thing I tried was Deletion on Counter column. Well, nothing happened. No error and the Counter column was sti

Re: How Cassandra determines the splits

2012-05-02 Thread Patrik Modesto
Hi, I had a simillar problem with Cassandra 0.8.x and the problem was when configured Cassandra with rpc_address: 0.0.0.0 and starting Hadoop job from outside the Cassandra cluster. But with version 1.0.x the problem is gone. You can debug the splits with thrift. This is a copy&paste part of my s

Re: AssertionError: originally calculated column size ...

2012-05-01 Thread Patrik Modesto
https://issues.apache.org/jira/browse/CASSANDRA-4206 Regards, Patrik On Tue, May 1, 2012 at 03:46, Jonathan Ellis wrote: > On Mon, Apr 30, 2012 at 2:11 PM, Patrik Modesto > wrote: >> I think the problem is somehow connected to an IntegerType secondary >> index. > > Coul

Re: AssertionError: originally calculated column size ...

2012-04-30 Thread Patrik Modesto
fault ? > > Cheers > > > - > Aaron Morton > Freelance Developer > @aaronmorton > http://www.thelastpickle.com > > On 27/04/2012, at 6:56 PM, Patrik Modesto wrote: > > Hi, > > I've 4 node cluster of Cassandra 1.0.9. There is a rfTest3 keyspa

Re: Proper Mutation to IntegerType column w/ secondary index

2012-04-25 Thread Patrik Modesto
I'll also try LongType index. I'll let you know how that works. Regards, P.

Proper Mutation to IntegerType column w/ secondary index

2012-04-25 Thread Patrik Modesto
Hi, I see strange behaviour using CF with two secondary indexes, one IntegerType one UTF8Type. Using Cassanra 1.0.9 and CF: create column family IndexTest with column_type=Standard and comparator=UTF8Type and default_validation_class=UTF8Type and key_validation_class=UTF8Type

Re: Poor write performance with seconrady index

2012-04-17 Thread Patrik Modesto
gt; > - > Aaron Morton > Freelance Developer > @aaronmorton > http://www.thelastpickle.com > > On 17/04/2012, at 2:51 AM, Patrik Modesto wrote: > > Hi, > > I've a 4 node test cluster running Cassandra 1.0.9, 32GB memory, 4x > 1TB disks. I&#x

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-08 Thread Patrik Modesto
dpoints'. > This method split token range of each node with describe_splits method, but > I think there is something wrong when you create Cassandra connection on > host '0.0.0.0'. > > > > > Le 7 mars 2012 09:07, Patrik Modesto a écrit : > >> You'

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-07 Thread Patrik Modesto
sure to see the good log file, I say that because in your first mail > you link the client job log. > And may be you can log the size of each split in CFIF. > > > > > Le 6 mars 2012 13:09, Patrik Modesto a écrit : > >> I've added a debug message in the CFRR.get

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
.org/jira/browse/CASSANDRA-3942 - I'm not sure if it > really affects the execution of the job itself though. > > On Mar 6, 2012, at 2:32 AM, Patrik Modesto wrote: > >> Hi, >> >> I was recently trying Hadoop job + cassandra-all 0.8.10 again and the >> Timeout

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
adoop estimate the task progress on 16384 inputs, it's why you >> have something like 9076.81%. >> >> If you can't change rpc_adress configuration, I don't know how you can >> solve your problem :/, sorry. >> >> Le 6 mars 2012 11:53, Patrik Modesto a

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
s between 0.8.7 and 0.8.10 > > > Le 6 mars 2012 09:32, Patrik Modesto a écrit : > >> Hi, >> >> I was recently trying Hadoop job + cassandra-all 0.8.10 again and the >> Timeouts I get are not because of the Cassandra can't handle the >> requests. I&#x

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
7_2027_m_73 9639.04% task_201202281457_2027_m_000105 10538.60% task_201202281457_2027_m_000108 9364.17% None of this happens with cassandra-all 0.8.7. Regards, P. On Tue, Feb 28, 2012 at 12:29, Patrik Modesto wrote: > I'll alter these settings and will let you know. > > Reg

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-28 Thread Patrik Modesto
s did not respond > in time. > > Cheers > > - > Aaron Morton > Freelance Developer > @aaronmorton > http://www.thelastpickle.com > > On 28/02/2012, at 8:18 PM, Patrik Modesto wrote: > > Hi aaron, > > this is our current settings: &

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-27 Thread Patrik Modesto
ave for cassandra.range.batch.size > and rpc_timeout_in_ms  ? Have you tried reducing the first and/or increasing > the second ? > > Cheers > > - > Aaron Morton > Freelance Developer > @aaronmorton > http://www.thelastpickle.com > > On 27/02/2012, a

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-26 Thread Patrik Modesto
On Sun, Feb 26, 2012 at 04:25, Edward Capriolo wrote: > Did you see the notes here? I'm not sure what do you mean by the notes? I'm using the mapred.* settings suggested there: mapred.max.tracker.failures 20 mapred.map.max.attempts 20

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
No, we have our own servers, no cloud service. P. On Feb 24, 2012 5:57 PM, "Jeremy Hanna" wrote: > By chance are you in EC2? > > On Feb 24, 2012, at 8:33 AM, Patrik Modesto wrote: > > > Hi Jeremy, > > > > I've seen the page and tried the values but

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
Hi Jeremy, I've seen the page and tried the values but to no help. Here goes tcpdump of one failed TCP connection: 15:06:20.231421 IP 10.0.18.87.9160 > 10.0.18.87.39396: Flags [P.], seq 137891735:137904068, ack 708, win 282, options [nop,nop,TS val 4119981636 ecr 4119981636], length 12333 15:06:

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
There are two more similar exceptions: DEBUG 14:23:12,817 Thrift transport error occurred during processing of message. org.apache.thrift.transport.TTransportException at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) at org.apache.thrift.transpor

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
Some more info, running the cluster with debug enabled shows this error: DEBUG 14:23:05,749 Thrift transport error occurred during processing of message. org.apache.thrift.transport.TTransportException: java.net.SocketException: Broken pipe at org.apache.thrift.transport.TIOStreamTranspor

newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
Hi, I can see some strange behaviour on my test cluster and in production. Both running cassandra 0.8.10. Strange is that when I compile my mapreduce job against cassandra-all 0.8.7 everything is ok, but if I use higher version I get quite a lots of TimedOutException. java.lang.RuntimeException:

Re: truncate command fails

2012-02-07 Thread Patrik Modesto
On Tue, Feb 7, 2012 at 18:35, Brandon Williams wrote: > Make sure you have JNA, without it the cost of forking ln to snapshot > is expensive. I assume JNA is enabled. >From the output.log: INFO 10:16:20,725 JNA mlockall successful Regards, P.

truncate command fails

2012-02-07 Thread Patrik Modesto
Hi, I've a strange problem with my test cluster. Calling truncate on a small ColumnFamily on idle cluster of 4 nodes returns UnavailableException after 10s. That 10s is set in rpc_timeout_in_ms. All nodes are up and running, here is nodetool ring: Address DC RackStatus S

Re: rpc_address: 0.0.0.0

2012-01-29 Thread Patrik Modesto
Thanks! I've created ticked https://issues.apache.org/jira/browse/CASSANDRA-3811 Regards, P. On Sun, Jan 29, 2012 at 20:00, Frisch, Michael wrote: > If the code in the 0.8 branch is reflective of what is actually included in > Cassandra 0.8.9 (here: > http://svn.apache.org/repos/asf/cassandra

rpc_address: 0.0.0.0

2012-01-26 Thread Patrik Modesto
Hi, #using cassandra 0.8.9 I used to have rpc_address set to 0.0.0.0 to bind cassandra to all interfaces. After upgrading out Hadoop jobs to cassandra 0.8.9 (from 0.8.7) there are lots of these messages, and the jobs fails. 12/01/26 11:15:21 DEBUG hadoop.ColumnFamilyInputFormat: failed connect

Re: Cassandra not suitable?

2011-12-07 Thread Patrik Modesto
ta dir? > > -Jake > > > On Tue, Dec 6, 2011 at 4:50 AM, Patrik Modesto > wrote: >> >> Hi, >> >> I'm quite desperate about Cassandra's performance in our production >> cluster. We have 8 real-HW nodes, 32core CPU, 32GB memory, 4 disks in &g

Re: Cassandra not suitable?

2011-12-07 Thread Patrik Modesto
d good > luck with MR/Pig over Cassandra, but it's after some lessons learned wrt > configuration of both Cassandra and Hadoop. > > On Dec 6, 2011, at 3:50 AM, Patrik Modesto wrote: > >> Hi, >> >> I'm quite desperate about Cassandra's performance in our

Cassandra not suitable?

2011-12-06 Thread Patrik Modesto
Hi, I'm quite desperate about Cassandra's performance in our production cluster. We have 8 real-HW nodes, 32core CPU, 32GB memory, 4 disks in raid10, cassandra 0.8.8, RF=3 and Hadoop. We four keyspaces, one is the large one, it has 2 CFs, one is kind of index, the other holds data. There are about

Re: Cassandra 0.8.8

2011-12-02 Thread Patrik Modesto
be early next week). > > -- > Sylvain > > On Fri, Dec 2, 2011 at 10:49 AM, Patrik Modesto > wrote: >> Hi, >> >> It's been almost 2 months since the release of the 0.8.7 version and >> there are quite some changes in 0.8.8, so I'd like to ask is there a >> release date? >> >> Regards, >> Patrik

Cassandra 0.8.8

2011-12-02 Thread Patrik Modesto
Hi, It's been almost 2 months since the release of the 0.8.7 version and there are quite some changes in 0.8.8, so I'd like to ask is there a release date? Regards, Patrik

Re: timeout while doing repair

2011-11-24 Thread Patrik Modesto
ds, Patrik Dne 24.11.2011 19:13 "Jahangir Mohammed" napsal(a): > What I know is timeout is because of increased load on node due to repair. > > Hardware? EC2? > > Did you check tpstats? > > On Thu, Nov 24, 2011 at 11:42 AM, Patrik Modesto > wrote: > >>

Re: timeout while doing repair

2011-11-24 Thread Patrik Modesto
timeout from client, look at rpc_timeout_in_ms. > Increase it and see if you still suffer this. > > Repair is a costly process. > > Thanks, > Jahangir Mohammed. > > > > On Thu, Nov 24, 2011 at 2:45 AM, Patrik Modesto > wrote: > >> Hi, >> >> I have a test cluste

timeout while doing repair

2011-11-23 Thread Patrik Modesto
Hi, I have a test cluster of 4 nodes running Debian and Cassandra 0.8.7, there are 3 keyspaces, all with RF=3, a node has load around 40GB. When I run "nodetool repair" after a while all thrift clients that read with CL.QUORUM get TimeoutException and even some that use just CL.ONE. I've tried to

decommissioned node still in "LoadMap" in JMX Management Console

2011-11-08 Thread Patrik Modesto
Hi, on our production cluster of 8 nodes which is running cassandra 0.8.7 we still see in the MBean "org.apache.cassandra.db:type=StorageService.LoadMap" in JMX Management console the 9th node we added for testing for a short time. After the testing we decommissioned the 9th node and has been rein

[SOLVED] Re: after upgrade to 0.8.7: InvalidRequestException(why:Column timestamp is required)

2011-10-20 Thread Patrik Modesto
As usually happen, I've found the problem just after I sent the question. I have to use setters for setting values to thrift.* classes. So instead of: Deletion d = new Deletion(); d.timestamp = 1; use: Deletion d = new Deletion(); d.setTimestamp(1); etc. Regards, Patrik

after upgrade to 0.8.7: InvalidRequestException(why:Column timestamp is required)

2011-10-20 Thread Patrik Modesto
Hi, I've a 4 node cluster of cassandra 0.8.7 (upgraded just recently from 0.7.8). Upgrade went smothly, no problem with the data. Problem are my MapReduce tasks. The all report: java.io.IOException: InvalidRequestException(why:Column timestamp is required) at org.apache.cassandra.hadoop

Re: RF=1 w/ hadoop jobs

2011-09-05 Thread Patrik Modesto
On Mon, Sep 5, 2011 at 09:39, Mick Semb Wever wrote: > I've entered a jira issue covering this request. > https://issues.apache.org/jira/browse/CASSANDRA-3136 > > Would you mind attaching your patch to the issue. > (No review of it will happen anywhere else.) I see Jonathan didn't change his mind

Re: RF=1 w/ hadoop jobs

2011-09-02 Thread Patrik Modesto
On Fri, Sep 2, 2011 at 08:54, Mick Semb Wever wrote: > Patrik: is it possible to describe the use-case you have here? Sure. We use Cassandra as a storage for web-pages, we store the HTML, all URLs that has the same HTML data and some computed data. We run Hadoop MR jobs to compute lexical and th

Re: RF=1 w/ hadoop jobs

2011-09-01 Thread Patrik Modesto
Hi, On Thu, Sep 1, 2011 at 12:36, Mck wrote: >> It's available here: http://pastebin.com/hhrr8m9P (for version 0.7.8) > > I'm interested in this patch and see it's usefulness but no one will act > until you attach it to an issue. (I think a new issue is appropriate > here). I'm glad someone is i

Re: RF=1

2011-08-19 Thread Patrik Modesto
Is there really no interest in the patch? P. On Thu, Aug 18, 2011 at 08:54, Patrik Modesto wrote: > On Wed, Aug 17, 2011 at 17:08, Jonathan Ellis wrote: >> See https://issues.apache.org/jira/browse/CASSANDRA-2388 > > Ok, thanks for the JIRA ticker. I've found that very same

Re: RF=1

2011-08-17 Thread Patrik Modesto
On Wed, Aug 17, 2011 at 17:08, Jonathan Ellis wrote: > See https://issues.apache.org/jira/browse/CASSANDRA-2388 Ok, thanks for the JIRA ticker. I've found that very same problem during my work on ignoring unavailabla ranges. But there is the another problem with Hadoop-Cassandra, if there is no

Re: RF=1

2011-08-17 Thread Patrik Modesto
And one more patch: http://pastebin.com/zfNPjtQz This one handles a case where there are no nodes available for a slice. For example where the is a keyspace with RF=1 and a node is shut down. Its range of keys gets ignored. Regards, P. On Wed, Aug 17, 2011 at 13:28, Patrik Modesto wrote: >

Re: RF=1

2011-08-17 Thread Patrik Modesto
gt; Cheers > - > Aaron Morton > Freelance Cassandra Developer > @aaronmorton > http://www.thelastpickle.com > > On 3 Aug 2011, at 16:18, Patrik Modesto wrote: > >> On Tue, Aug 2, 2011 at 23:10, Jeremiah Jordan >> wrote: >>> If you have RF=1, t

Re: Internal error processing get_range_slices

2011-08-15 Thread Patrik Modesto
On Mon, Aug 15, 2011 at 15:09, Jonathan Ellis wrote: > On Mon, Aug 15, 2011 at 7:13 AM, Patrik Modesto > wrote: >> PS: while reading the email before I'd send it, I've noticed the >> keyRange.count =... is it possible that Cassandra is preallocating >&g

Internal error processing get_range_slices

2011-08-15 Thread Patrik Modesto
Hi, on our dev cluster of 4 cassandra nodes 0.7.8 I'm suddenly getting: ERROR 13:40:50,848 Internal error processing get_range_slices java.lang.OutOfMemoryError: Java heap space at java.util.ArrayList.(ArrayList.java:112) at org.apache.cassandra.service.StorageProxy.getRangeSlice

Re: Compacting large row

2011-08-06 Thread Patrik Modesto
On Fri, Aug 5, 2011 at 15:02, Jonathan Ellis wrote: > It's logging the actual key, not the md5.  It's just converting the > key bytes to hex first to make sure it's printable. Great! I'm using MD5 as a key so I didn't notice that. Thanks, P.

Compacting large row

2011-08-05 Thread Patrik Modesto
Hi, I got several "Compacting large row ..." messages while runinng a compaction. Is there any way to get the row to see it? I know it's MD5 but it's MD5 of a key and I need the key to get to the row. The sstable is really huge so dump is not an option. Using Cassandra 0.7.8 Regards, P.

Re: RF=1

2011-08-02 Thread Patrik Modesto
On Tue, Aug 2, 2011 at 23:10, Jeremiah Jordan wrote: > If you have RF=1, taking one node down is going to cause 25% of your > data to be unavailable.  If you want to tolerate a machines going down > you need to have at least RF=2, if you want to use quorum and have a > machine go down, you need at

RF=1

2011-08-02 Thread Patrik Modesto
Hi all! I've a test cluster of 4 nodes running cassandra 0.7.8, with one keyspace with RF=1, each node owns 25% of the data. As long as all nodes are alive, there is no problem, but when I shut down just one node I get UnavailableException in my application. cassandra-cli returns "null" and hadoop

A way to break the cluster

2011-03-15 Thread Patrik Modesto
Hi, I'm did break my test cluster again. It's really strange. I use cassandra 0.7.3. This is what I did: - install node1 - install node2, auto_bootstrap: true - install node3, auto_bootstrap: true - created a keyspace with RF 1, populate with data - create a keyspace with RF 3, populate with dat

Re: problem with bootstrap

2011-03-14 Thread Patrik Modesto
On Fri, Mar 11, 2011 at 22:31, Aaron Morton wrote: > The assertion is interesting. Can you reproduce it with logging at debug and > post the results? Could you try to reproduce it with a clean cluster? It was on a clean cluster last time. Anyway I started clean cluster again, repeated the same s

Re: problem with bootstrap

2011-03-11 Thread Patrik Modesto
Unfortunately I can't provide the info, I deleted it. It was in wery strange state. I started with new cluster today, 2 nodes, each with auto_bootstrap:true. I can create a keyspace with RF=3, but I can't insert any data in it. It didn't happen with the old cluster which made me think. How could I

Re: problem with bootstrap

2011-03-10 Thread Patrik Modesto
Hi, I'm stil fighting the Exception in thread "main" java.lang.IllegalStateException: replication factor (3) exceeds number of endpoints (2). When I have a 2-server cluster, create Keyspace with RF 3, I'm able to add (without auto_bootstrap) another node but cluster nodetool commands don't work a

problem with bootstrap

2011-03-08 Thread Patrik Modesto
Hi, I've small test cluster, 2 servers, both running successfully cassandra 0.7.3. I've three keyspaces, two with RF1, one with RF3. Now when I try to bootstrap 3rd server (empty initial_token, auto_bootstrap: true), I get this exception on the new server. INFO 23:13:43,229 Joining: getting boot

Column name size

2011-02-11 Thread Patrik Modesto
Hi all! I'm thinking if size of a column name could matter for a large dataset in Cassandra (I mean lots of rows). For example what if I have a row with 10 columns each has 10 bytes value and 10 bytes name. Do I have half the row size just of the column names and the other half of the data (not c

Re: OOM during batch_mutate

2011-02-08 Thread Patrik Modesto
On Tue, Feb 8, 2011 at 00:05, Jonathan Ellis wrote: > Sounds like the keyspace was created on the 32GB machine, so it > guessed memtable sizes that are too large when run on the 16GB one. > Use "update column family" from the cli to cut the throughput and > operations thresholds in half, or to 1/4

Re: [0.7.1] Error in ThreadPoolExecutor

2011-02-07 Thread Patrik Modesto
Hi, here is the ticket: https://issues.apache.org/jira/browse/CASSANDRA-2134 I'm using the default partitioner, that should be the RandomPartitioner. HTH, Patrik On Tue, Feb 8, 2011 at 00:03, Jonathan Ellis wrote: > Can you open a ticket for this?  And are you using order-preserving > partit

Re: [0.7.1] more exceptions: Illegal mode

2011-02-07 Thread Patrik Modesto
On Mon, Feb 7, 2011 at 15:42, Thibaut Britz wrote: > I think this is related to a faulty disk. I'm not sure thats the problem. Cassandra 0.7.0 didn't report any problem. It started with Cassandra 0.7.1. Patrik

Re: OOM during batch_mutate

2011-02-07 Thread Patrik Modesto
On Mon, Feb 7, 2011 at 15:44, sridhar basam wrote: > Looks like you don't have a big enough working set from your GC logs, there > doesn't seem to be a lot being reclaimed in the GC process. The process is > reclaiming a few hundred MB and is running every few seconds. How big are > your caches? T

[0.7.1] more exceptions: Illegal mode

2011-02-07 Thread Patrik Modesto
INFO 15:30:49,647 Compacted to /www/foo/cassandra/data/foo/Url-tmp-f-767-Data.db. 4,199,999,762 to 4,162,579,242 (~99% of original) bytes for 379,179 keys. Time: 137,149ms. ERROR 15:30:49,699 Fatal exception in thread Thread[CompactionExecutor:1,1,main] java.lang.RuntimeException: java.lang.Illeg

Re: OOM during batch_mutate

2011-02-07 Thread Patrik Modesto
Just tried current 0.7.1 from cassandra-0.7 branch and it does the same. OOM after three runs. -Xm* setting is computed by cassandra-env.sh like this: -Xms8022M -Xmx8022M -Xmn2005M What am I doing wrong? Thanks, Patrik On Mon, Feb 7, 2011 at 14:18, Patrik Modesto wrote: > I forgot to ment

Re: OOM during batch_mutate

2011-02-07 Thread Patrik Modesto
I forgot to mention I use 0.7.0 stable version. HTH, Patrik

[0.7.1] Error in ThreadPoolExecutor

2011-02-07 Thread Patrik Modesto
Hi, on my two-node test setup I get repeatedly following error: The 10.0.18.129 server log: INFO 14:10:37,707 Node /10.0.18.99 has restarted, now UP again INFO 14:10:37,708 Checking remote schema before delivering hints INFO 14:10:37,708 Sleeping 45506ms to stagger hint delivery INFO 14:10:3

OOM during batch_mutate

2011-02-07 Thread Patrik Modesto
Hi all! I'm running into OOM problem during batch_mutate. I've a test cluster of two servers, 32GB and 16GB RAM, real HW. I've one keyspace and one CF with 1,4mil rows, each 10 columns. A row is around 5k in size. I run Hadoop MR task that reads one column and generates Mutation that updates anoth

get_range_slices and tombstones

2011-02-04 Thread Patrik Modesto
Hi! I'm getting tombstones from get_range_slices(). I know that's normal. But is there a way to know that a key is tombstone? I know tombstone has no columns but I can create a row without any columns that would look like a tombstone in get_range_slices(). Regards, Patrik

Re: [mapreduce] ColumnFamilyRecordWriter hidden reuse

2011-01-26 Thread Patrik Modesto
On Wed, Jan 26, 2011 at 08:58, Mck wrote: >> You are correct that microseconds would be better but for the test it >> doesn't matter that much. > > Have you tried. I'm very new to cassandra as well, and always uncertain > as to what to expect... IMHO it's matter of use-case. In my use-case there

Re: [mapreduce] ColumnFamilyRecordWriter hidden reuse

2011-01-25 Thread Patrik Modesto
On Tue, Jan 25, 2011 at 19:09, Mick Semb Wever wrote: > In fact i have another problem (trying to write an empty byte[], or > something, as a key, which put one whole row out of whack, ((one row in > 25 million...))). > > But i'm debugging along the same code. > > I don't quite understand how the

Re: [mapreduce] ColumnFamilyRecordWriter hidden reuse

2011-01-25 Thread Patrik Modesto
Hi Mick, attached is the very simple MR job, that deletes expired URL from my test Cassandra DB. The keyspace looks like this: Keyspace: Test: Replication Strategy: org.apache.cassandra.locator.SimpleStrategy Replication Factor: 2 Column Families: ColumnFamily: Url2 Columns sort

[mapreduce] ColumnFamilyRecordWriter hidden reuse

2011-01-25 Thread Patrik Modesto
Hi, I play with Cassandra 0.7.0 and Hadoop, developing simple MapReduce tasks. While developing really simple MR task, I've found that a combiantion of Hadoop optimalization and Cassandra ColumnFamilyRecordWriter queue creates wrong keys to send to batch_mutate(). The proble is in the reduce part,

Re: WordCount example problem

2010-11-10 Thread Patrik Modesto
That's exactly what's happening to me. I wonder why Google did't find it. Thanks! P. On Wed, Nov 10, 2010 at 15:39, Jonathan Ellis wrote: > http://www.mail-archive.com/user@cassandra.apache.org/msg07093.html > > On Wed, Nov 10, 2010 at 5:47 AM, Patrik Modesto > wro

Re: WordCount example problem

2010-11-10 Thread Patrik Modesto
e- > From: Jonathan Ellis [mailto:jbel...@gmail.com] > Sent: Wednesday, November 10, 2010 8:40 AM > To: user > Subject: Re: WordCount example problem > > http://www.mail-archive.com/user@cassandra.apache.org/msg07093.html > > On Wed, Nov 10, 2010 at 5:47 AM, Patrik Mo

Re: WordCount example problem

2010-11-10 Thread Patrik Modesto
Hi, I'm trying the WordCount example and getting this error: [12:33]$ ./bin/word_count 10/11/10 12:34:35 INFO WordCount: output reducer type: filesystem 10/11/10 12:34:36 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId= 10/11/10 12:34:36 INFO WordCount: XXX:te