It works now, thanks!
P.
On Wed, Apr 24, 2013 at 2:53 PM, Sylvain Lebresne wrote:
> Sorry, seems I screwed up somehow.
> That should be fixed now however.
Hi!
I can't find the 1.1.11 package for the Debian at
http://www.apache.org/dist/cassandra/debian/
The .deb package is there but Packages files still contain just version 1.1.10.
Regards,
Patrik
On Fri, Jul 20, 2012 at 11:17 AM, aaron morton wrote:
> Ordering the rows by row key locally would mean that every row on the node
> would have to be scanned to find the ones whose token was in the required
> token range.
I don't know much about Cassandra internals, but from a user point of
view,
ote:
> The probem could be while redistributing the tokens. In that case the
> hashes has to be recalculated on each fo the candidate node.
>
> -Thanks,
> Prasenjit
>
> On Thu, Jul 19, 2012 at 12:19 PM, Patrik Modesto
> wrote:
>> Hi,
>>
>> I know that Ran
Hi,
I know that RandomPartitioner does MD5 of a key and the MD5 is then
used for key distribution AND key ordering. I was just wondering if
it's possible to have RandomPartitioner just for key distribution and
OrderedPartitioner just for per-node key ordering. That would solve
the often requested
On Mon, Jun 25, 2012 at 9:28 AM, Sylvain Lebresne wrote:
> On Mon, Jun 25, 2012 at 9:06 AM, Patrik Modesto
> wrote:
>> I'm used to use Mutation for everything, so the first thing I tried
>> was Deletion on Counter column. Well, nothing happened. No error and
>> the C
Hi,
I was playing with Counters in Cassandra 1.0.10 and I see a behaviour
that I didn't expect. It's about removing a Counter column.
I'm used to use Mutation for everything, so the first thing I tried
was Deletion on Counter column. Well, nothing happened. No error and
the Counter column was sti
Hi,
I had a simillar problem with Cassandra 0.8.x and the problem was when
configured Cassandra with rpc_address: 0.0.0.0 and starting Hadoop job
from outside the Cassandra cluster. But with version 1.0.x the problem
is gone.
You can debug the splits with thrift. This is a copy&paste part of my
s
https://issues.apache.org/jira/browse/CASSANDRA-4206
Regards,
Patrik
On Tue, May 1, 2012 at 03:46, Jonathan Ellis wrote:
> On Mon, Apr 30, 2012 at 2:11 PM, Patrik Modesto
> wrote:
>> I think the problem is somehow connected to an IntegerType secondary
>> index.
>
> Coul
fault ?
>
> Cheers
>
>
> -
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 27/04/2012, at 6:56 PM, Patrik Modesto wrote:
>
> Hi,
>
> I've 4 node cluster of Cassandra 1.0.9. There is a rfTest3 keyspa
I'll also try LongType index. I'll let you know how that works.
Regards,
P.
Hi,
I see strange behaviour using CF with two secondary indexes, one
IntegerType one UTF8Type.
Using Cassanra 1.0.9 and CF:
create column family IndexTest
with column_type=Standard
and comparator=UTF8Type
and default_validation_class=UTF8Type
and key_validation_class=UTF8Type
gt;
> -
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 17/04/2012, at 2:51 AM, Patrik Modesto wrote:
>
> Hi,
>
> I've a 4 node test cluster running Cassandra 1.0.9, 32GB memory, 4x
> 1TB disks. I
dpoints'.
> This method split token range of each node with describe_splits method, but
> I think there is something wrong when you create Cassandra connection on
> host '0.0.0.0'.
>
>
>
>
> Le 7 mars 2012 09:07, Patrik Modesto a écrit :
>
>> You'
sure to see the good log file, I say that because in your first mail
> you link the client job log.
> And may be you can log the size of each split in CFIF.
>
>
>
>
> Le 6 mars 2012 13:09, Patrik Modesto a écrit :
>
>> I've added a debug message in the CFRR.get
.org/jira/browse/CASSANDRA-3942 - I'm not sure if it
> really affects the execution of the job itself though.
>
> On Mar 6, 2012, at 2:32 AM, Patrik Modesto wrote:
>
>> Hi,
>>
>> I was recently trying Hadoop job + cassandra-all 0.8.10 again and the
>> Timeout
adoop estimate the task progress on 16384 inputs, it's why you
>> have something like 9076.81%.
>>
>> If you can't change rpc_adress configuration, I don't know how you can
>> solve your problem :/, sorry.
>>
>> Le 6 mars 2012 11:53, Patrik Modesto a
s between 0.8.7 and 0.8.10
>
>
> Le 6 mars 2012 09:32, Patrik Modesto a écrit :
>
>> Hi,
>>
>> I was recently trying Hadoop job + cassandra-all 0.8.10 again and the
>> Timeouts I get are not because of the Cassandra can't handle the
>> requests. I
7_2027_m_73 9639.04%
task_201202281457_2027_m_000105 10538.60%
task_201202281457_2027_m_000108 9364.17%
None of this happens with cassandra-all 0.8.7.
Regards,
P.
On Tue, Feb 28, 2012 at 12:29, Patrik Modesto wrote:
> I'll alter these settings and will let you know.
>
> Reg
s did not respond
> in time.
>
> Cheers
>
> -
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 28/02/2012, at 8:18 PM, Patrik Modesto wrote:
>
> Hi aaron,
>
> this is our current settings:
&
ave for cassandra.range.batch.size
> and rpc_timeout_in_ms ? Have you tried reducing the first and/or increasing
> the second ?
>
> Cheers
>
> -
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 27/02/2012, a
On Sun, Feb 26, 2012 at 04:25, Edward Capriolo wrote:
> Did you see the notes here?
I'm not sure what do you mean by the notes?
I'm using the mapred.* settings suggested there:
mapred.max.tracker.failures
20
mapred.map.max.attempts
20
No, we have our own servers, no cloud service.
P.
On Feb 24, 2012 5:57 PM, "Jeremy Hanna" wrote:
> By chance are you in EC2?
>
> On Feb 24, 2012, at 8:33 AM, Patrik Modesto wrote:
>
> > Hi Jeremy,
> >
> > I've seen the page and tried the values but
Hi Jeremy,
I've seen the page and tried the values but to no help.
Here goes tcpdump of one failed TCP connection:
15:06:20.231421 IP 10.0.18.87.9160 > 10.0.18.87.39396: Flags [P.], seq
137891735:137904068, ack 708, win 282, options [nop,nop,TS val
4119981636 ecr 4119981636], length 12333
15:06:
There are two more similar exceptions:
DEBUG 14:23:12,817 Thrift transport error occurred during processing of message.
org.apache.thrift.transport.TTransportException
at
org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
at org.apache.thrift.transpor
Some more info, running the cluster with debug enabled shows this error:
DEBUG 14:23:05,749 Thrift transport error occurred during processing of message.
org.apache.thrift.transport.TTransportException:
java.net.SocketException: Broken pipe
at
org.apache.thrift.transport.TIOStreamTranspor
Hi,
I can see some strange behaviour on my test cluster and in production.
Both running cassandra 0.8.10. Strange is that when I compile my
mapreduce job against cassandra-all 0.8.7 everything is ok, but if I
use higher version I get quite a lots of TimedOutException.
java.lang.RuntimeException:
On Tue, Feb 7, 2012 at 18:35, Brandon Williams wrote:
> Make sure you have JNA, without it the cost of forking ln to snapshot
> is expensive.
I assume JNA is enabled.
>From the output.log:
INFO 10:16:20,725 JNA mlockall successful
Regards,
P.
Hi,
I've a strange problem with my test cluster. Calling truncate on a
small ColumnFamily on idle cluster of 4 nodes returns
UnavailableException after 10s. That 10s is set in rpc_timeout_in_ms.
All nodes are up and running, here is nodetool ring:
Address DC RackStatus S
Thanks! I've created ticked https://issues.apache.org/jira/browse/CASSANDRA-3811
Regards,
P.
On Sun, Jan 29, 2012 at 20:00, Frisch, Michael
wrote:
> If the code in the 0.8 branch is reflective of what is actually included in
> Cassandra 0.8.9 (here:
> http://svn.apache.org/repos/asf/cassandra
Hi,
#using cassandra 0.8.9
I used to have rpc_address set to 0.0.0.0 to bind cassandra to all
interfaces. After upgrading out Hadoop jobs to cassandra 0.8.9 (from
0.8.7) there are lots of these messages, and the jobs fails.
12/01/26 11:15:21 DEBUG hadoop.ColumnFamilyInputFormat: failed
connect
ta dir?
>
> -Jake
>
>
> On Tue, Dec 6, 2011 at 4:50 AM, Patrik Modesto
> wrote:
>>
>> Hi,
>>
>> I'm quite desperate about Cassandra's performance in our production
>> cluster. We have 8 real-HW nodes, 32core CPU, 32GB memory, 4 disks in
&g
d good
> luck with MR/Pig over Cassandra, but it's after some lessons learned wrt
> configuration of both Cassandra and Hadoop.
>
> On Dec 6, 2011, at 3:50 AM, Patrik Modesto wrote:
>
>> Hi,
>>
>> I'm quite desperate about Cassandra's performance in our
Hi,
I'm quite desperate about Cassandra's performance in our production
cluster. We have 8 real-HW nodes, 32core CPU, 32GB memory, 4 disks in
raid10, cassandra 0.8.8, RF=3 and Hadoop.
We four keyspaces, one is the large one, it has 2 CFs, one is kind of
index, the other holds data. There are about
be early next week).
>
> --
> Sylvain
>
> On Fri, Dec 2, 2011 at 10:49 AM, Patrik Modesto
> wrote:
>> Hi,
>>
>> It's been almost 2 months since the release of the 0.8.7 version and
>> there are quite some changes in 0.8.8, so I'd like to ask is there a
>> release date?
>>
>> Regards,
>> Patrik
Hi,
It's been almost 2 months since the release of the 0.8.7 version and
there are quite some changes in 0.8.8, so I'd like to ask is there a
release date?
Regards,
Patrik
ds,
Patrik
Dne 24.11.2011 19:13 "Jahangir Mohammed"
napsal(a):
> What I know is timeout is because of increased load on node due to repair.
>
> Hardware? EC2?
>
> Did you check tpstats?
>
> On Thu, Nov 24, 2011 at 11:42 AM, Patrik Modesto > wrote:
>
>>
timeout from client, look at rpc_timeout_in_ms.
> Increase it and see if you still suffer this.
>
> Repair is a costly process.
>
> Thanks,
> Jahangir Mohammed.
>
>
>
> On Thu, Nov 24, 2011 at 2:45 AM, Patrik Modesto
> wrote:
>
>> Hi,
>>
>> I have a test cluste
Hi,
I have a test cluster of 4 nodes running Debian and Cassandra 0.8.7,
there are 3 keyspaces, all with RF=3, a node has load around 40GB.
When I run "nodetool repair" after a while all thrift clients that
read with CL.QUORUM get TimeoutException and even some that use just
CL.ONE. I've tried to
Hi,
on our production cluster of 8 nodes which is running cassandra 0.8.7
we still see in the MBean
"org.apache.cassandra.db:type=StorageService.LoadMap" in JMX
Management console the 9th node we added for testing for a short time.
After the testing we decommissioned the 9th node and has been
rein
As usually happen, I've found the problem just after I sent the
question. I have to use setters for setting values to thrift.*
classes.
So instead of:
Deletion d = new Deletion();
d.timestamp = 1;
use:
Deletion d = new Deletion();
d.setTimestamp(1);
etc.
Regards,
Patrik
Hi,
I've a 4 node cluster of cassandra 0.8.7 (upgraded just recently from
0.7.8). Upgrade went smothly, no problem with the data.
Problem are my MapReduce tasks. The all report:
java.io.IOException: InvalidRequestException(why:Column timestamp is required)
at
org.apache.cassandra.hadoop
On Mon, Sep 5, 2011 at 09:39, Mick Semb Wever wrote:
> I've entered a jira issue covering this request.
> https://issues.apache.org/jira/browse/CASSANDRA-3136
>
> Would you mind attaching your patch to the issue.
> (No review of it will happen anywhere else.)
I see Jonathan didn't change his mind
On Fri, Sep 2, 2011 at 08:54, Mick Semb Wever wrote:
> Patrik: is it possible to describe the use-case you have here?
Sure.
We use Cassandra as a storage for web-pages, we store the HTML, all
URLs that has the same HTML data and some computed data. We run Hadoop
MR jobs to compute lexical and th
Hi,
On Thu, Sep 1, 2011 at 12:36, Mck wrote:
>> It's available here: http://pastebin.com/hhrr8m9P (for version 0.7.8)
>
> I'm interested in this patch and see it's usefulness but no one will act
> until you attach it to an issue. (I think a new issue is appropriate
> here).
I'm glad someone is i
Is there really no interest in the patch?
P.
On Thu, Aug 18, 2011 at 08:54, Patrik Modesto wrote:
> On Wed, Aug 17, 2011 at 17:08, Jonathan Ellis wrote:
>> See https://issues.apache.org/jira/browse/CASSANDRA-2388
>
> Ok, thanks for the JIRA ticker. I've found that very same
On Wed, Aug 17, 2011 at 17:08, Jonathan Ellis wrote:
> See https://issues.apache.org/jira/browse/CASSANDRA-2388
Ok, thanks for the JIRA ticker. I've found that very same problem
during my work on ignoring unavailabla ranges.
But there is the another problem with Hadoop-Cassandra, if there is no
And one more patch: http://pastebin.com/zfNPjtQz
This one handles a case where there are no nodes available for a
slice. For example where the is a keyspace with RF=1 and a node is
shut down. Its range of keys gets ignored.
Regards,
P.
On Wed, Aug 17, 2011 at 13:28, Patrik Modesto wrote:
>
gt; Cheers
> -
> Aaron Morton
> Freelance Cassandra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 3 Aug 2011, at 16:18, Patrik Modesto wrote:
>
>> On Tue, Aug 2, 2011 at 23:10, Jeremiah Jordan
>> wrote:
>>> If you have RF=1, t
On Mon, Aug 15, 2011 at 15:09, Jonathan Ellis wrote:
> On Mon, Aug 15, 2011 at 7:13 AM, Patrik Modesto
> wrote:
>> PS: while reading the email before I'd send it, I've noticed the
>> keyRange.count =... is it possible that Cassandra is preallocating
>&g
Hi,
on our dev cluster of 4 cassandra nodes 0.7.8 I'm suddenly getting:
ERROR 13:40:50,848 Internal error processing get_range_slices
java.lang.OutOfMemoryError: Java heap space
at java.util.ArrayList.(ArrayList.java:112)
at
org.apache.cassandra.service.StorageProxy.getRangeSlice
On Fri, Aug 5, 2011 at 15:02, Jonathan Ellis wrote:
> It's logging the actual key, not the md5. It's just converting the
> key bytes to hex first to make sure it's printable.
Great! I'm using MD5 as a key so I didn't notice that.
Thanks,
P.
Hi,
I got several "Compacting large row ..." messages while runinng a
compaction. Is there any way to get the row to see it? I know it's MD5
but it's MD5 of a key and I need the key to get to the row. The
sstable is really huge so dump is not an option.
Using Cassandra 0.7.8
Regards,
P.
On Tue, Aug 2, 2011 at 23:10, Jeremiah Jordan
wrote:
> If you have RF=1, taking one node down is going to cause 25% of your
> data to be unavailable. If you want to tolerate a machines going down
> you need to have at least RF=2, if you want to use quorum and have a
> machine go down, you need at
Hi all!
I've a test cluster of 4 nodes running cassandra 0.7.8, with one
keyspace with RF=1, each node owns 25% of the data. As long as all
nodes are alive, there is no problem, but when I shut down just one
node I get UnavailableException in my application. cassandra-cli
returns "null" and hadoop
Hi,
I'm did break my test cluster again. It's really strange. I use
cassandra 0.7.3. This is what I did:
- install node1
- install node2, auto_bootstrap: true
- install node3, auto_bootstrap: true
- created a keyspace with RF 1, populate with data
- create a keyspace with RF 3, populate with dat
On Fri, Mar 11, 2011 at 22:31, Aaron Morton wrote:
> The assertion is interesting. Can you reproduce it with logging at debug and
> post the results? Could you try to reproduce it with a clean cluster?
It was on a clean cluster last time. Anyway I started clean cluster
again, repeated the same s
Unfortunately I can't provide the info, I deleted it. It was in wery
strange state.
I started with new cluster today, 2 nodes, each with
auto_bootstrap:true. I can create a keyspace with RF=3, but I can't
insert any data in it. It didn't happen with the old cluster which
made me think. How could I
Hi,
I'm stil fighting the
Exception in thread "main" java.lang.IllegalStateException:
replication factor (3) exceeds number of endpoints (2).
When I have a 2-server cluster, create Keyspace with RF 3, I'm able to
add (without auto_bootstrap) another node but cluster nodetool
commands don't work a
Hi,
I've small test cluster, 2 servers, both running successfully
cassandra 0.7.3. I've three keyspaces, two with RF1, one with RF3. Now
when I try to bootstrap 3rd server (empty initial_token,
auto_bootstrap: true), I get this exception on the new server.
INFO 23:13:43,229 Joining: getting boot
Hi all!
I'm thinking if size of a column name could matter for a large dataset
in Cassandra (I mean lots of rows). For example what if I have a row
with 10 columns each has 10 bytes value and 10 bytes name. Do I have
half the row size just of the column names and the other half of the
data (not c
On Tue, Feb 8, 2011 at 00:05, Jonathan Ellis wrote:
> Sounds like the keyspace was created on the 32GB machine, so it
> guessed memtable sizes that are too large when run on the 16GB one.
> Use "update column family" from the cli to cut the throughput and
> operations thresholds in half, or to 1/4
Hi,
here is the ticket: https://issues.apache.org/jira/browse/CASSANDRA-2134
I'm using the default partitioner, that should be the RandomPartitioner.
HTH,
Patrik
On Tue, Feb 8, 2011 at 00:03, Jonathan Ellis wrote:
> Can you open a ticket for this? And are you using order-preserving
> partit
On Mon, Feb 7, 2011 at 15:42, Thibaut Britz
wrote:
> I think this is related to a faulty disk.
I'm not sure thats the problem. Cassandra 0.7.0 didn't report any
problem. It started with Cassandra 0.7.1.
Patrik
On Mon, Feb 7, 2011 at 15:44, sridhar basam wrote:
> Looks like you don't have a big enough working set from your GC logs, there
> doesn't seem to be a lot being reclaimed in the GC process. The process is
> reclaiming a few hundred MB and is running every few seconds. How big are
> your caches? T
INFO 15:30:49,647 Compacted to
/www/foo/cassandra/data/foo/Url-tmp-f-767-Data.db. 4,199,999,762 to
4,162,579,242 (~99% of original) bytes for 379,179 keys. Time:
137,149ms.
ERROR 15:30:49,699 Fatal exception in thread Thread[CompactionExecutor:1,1,main]
java.lang.RuntimeException: java.lang.Illeg
Just tried current 0.7.1 from cassandra-0.7 branch and it does the
same. OOM after three runs.
-Xm* setting is computed by cassandra-env.sh like this: -Xms8022M
-Xmx8022M -Xmn2005M
What am I doing wrong?
Thanks,
Patrik
On Mon, Feb 7, 2011 at 14:18, Patrik Modesto wrote:
> I forgot to ment
I forgot to mention I use 0.7.0 stable version.
HTH,
Patrik
Hi,
on my two-node test setup I get repeatedly following error:
The 10.0.18.129 server log:
INFO 14:10:37,707 Node /10.0.18.99 has restarted, now UP again
INFO 14:10:37,708 Checking remote schema before delivering hints
INFO 14:10:37,708 Sleeping 45506ms to stagger hint delivery
INFO 14:10:3
Hi all!
I'm running into OOM problem during batch_mutate. I've a test cluster
of two servers, 32GB and 16GB RAM, real HW. I've one keyspace and one
CF with 1,4mil rows, each 10 columns. A row is around 5k in size. I
run Hadoop MR task that reads one column and generates Mutation that
updates anoth
Hi!
I'm getting tombstones from get_range_slices(). I know that's normal.
But is there a way to know that a key is tombstone? I know tombstone
has no columns but I can create a row without any columns that would
look like a tombstone in get_range_slices().
Regards,
Patrik
On Wed, Jan 26, 2011 at 08:58, Mck wrote:
>> You are correct that microseconds would be better but for the test it
>> doesn't matter that much.
>
> Have you tried. I'm very new to cassandra as well, and always uncertain
> as to what to expect...
IMHO it's matter of use-case. In my use-case there
On Tue, Jan 25, 2011 at 19:09, Mick Semb Wever wrote:
> In fact i have another problem (trying to write an empty byte[], or
> something, as a key, which put one whole row out of whack, ((one row in
> 25 million...))).
>
> But i'm debugging along the same code.
>
> I don't quite understand how the
Hi Mick,
attached is the very simple MR job, that deletes expired URL from my
test Cassandra DB. The keyspace looks like this:
Keyspace: Test:
Replication Strategy: org.apache.cassandra.locator.SimpleStrategy
Replication Factor: 2
Column Families:
ColumnFamily: Url2
Columns sort
Hi,
I play with Cassandra 0.7.0 and Hadoop, developing simple MapReduce
tasks. While developing really simple MR task, I've found that a
combiantion of Hadoop optimalization and Cassandra
ColumnFamilyRecordWriter queue creates wrong keys to send to
batch_mutate(). The proble is in the reduce part,
That's exactly what's happening to me. I wonder why Google did't find it.
Thanks!
P.
On Wed, Nov 10, 2010 at 15:39, Jonathan Ellis wrote:
> http://www.mail-archive.com/user@cassandra.apache.org/msg07093.html
>
> On Wed, Nov 10, 2010 at 5:47 AM, Patrik Modesto
> wro
e-
> From: Jonathan Ellis [mailto:jbel...@gmail.com]
> Sent: Wednesday, November 10, 2010 8:40 AM
> To: user
> Subject: Re: WordCount example problem
>
> http://www.mail-archive.com/user@cassandra.apache.org/msg07093.html
>
> On Wed, Nov 10, 2010 at 5:47 AM, Patrik Mo
Hi,
I'm trying the WordCount example and getting this error:
[12:33]$ ./bin/word_count
10/11/10 12:34:35 INFO WordCount: output reducer type: filesystem
10/11/10 12:34:36 INFO jvm.JvmMetrics: Initializing JVM Metrics with
processName=JobTracker, sessionId=
10/11/10 12:34:36 INFO WordCount: XXX:te
78 matches
Mail list logo