is there hector snapshot available for 1.0rc? Hector 0.8 do not works
with it
Are you on a current Cassandra release? If not you could be seeing
https://issues.apache.org/jira/browse/CASSANDRA-2870
On Thu, Sep 29, 2011 at 7:02 PM, Anthony Ikeda
wrote:
> We seem to be having issues with out Consistency policies. We have it
> configured in Spring using the following:
>
>
Which version are you using? In my memory 0.8.3 cannot do it correctly but
later versions fixed the bug.
從我的 BlackBerry® 無線裝置
-Original Message-
From: Ramesh S
Date: Thu, 29 Sep 2011 15:23:29
To:
Reply-To: user@cassandra.apache.org
Subject: create super column family for
I am trying
Well I did a local test by debugging our code and the policy is set to
LOCAL_QUORUM. Is there something else I'm missing? More info that would
help? Or should I direct this to the hector-users group?
Anthony
On Thu, Sep 29, 2011 at 5:13 PM, Anthony Ikeda
wrote:
> RF=3
>
> 2 DC's
> 3 nodes each
RF=3
2 DC's
3 nodes each
On Thu, Sep 29, 2011 at 5:08 PM, Paul Loy wrote:
> What is your replication factor?
>
>
> On Thu, Sep 29, 2011 at 5:02 PM, Anthony Ikeda <
> anthony.ikeda@gmail.com> wrote:
>
>> We seem to be having issues with out Consistency policies. We have it
>> configured in
What is your replication factor?
On Thu, Sep 29, 2011 at 5:02 PM, Anthony Ikeda
wrote:
> We seem to be having issues with out Consistency policies. We have it
> configured in Spring using the following:
>
> "me.prettyprint.cassandra.model.ConfigurableConsistencyLevel">
>
>
>
>
>
>
We seem to be having issues with out Consistency policies. We have it
configured in Spring using the following:
However, in our distributed testing, bringing down a single node will cause
the 'May not be enough replicas present to
right. path == null means "Cassandra is out of disk space."
On Thu, Sep 29, 2011 at 5:06 PM, Yang wrote:
> I am using mmap:
>
> INFO [main] 2011-09-29 21:35:21,458 DatabaseDescriptor.java (line
> 177) DiskAccessMode is mmap, indexAccessMode is mmap
>
>
>
> I think the cleanerMethod must be OK
I am using mmap:
INFO [main] 2011-09-29 21:35:21,458 DatabaseDescriptor.java (line
177) DiskAccessMode is mmap, indexAccessMode is mmap
I think the cleanerMethod must be OK now,
so the condition for GC
if (path == null
&& (DatabaseDescriptor.getDiskAccessMode() ==
Config.Di
Thanks Yi,
I am using compound rowkeys in other column families which use this Id (UUID
in this case). What I am looking for is unique Ids as a base point and I am
hoping UUID is the best way to go.
regards,
Ramesh
On Thu, Sep 29, 2011 at 4:28 PM, Yi Yang wrote:
> ** I don't know if I unders
you're not using mmap i/o mode?
On Thu, Sep 29, 2011 at 2:21 PM, Yang wrote:
> I see now :
>
> https://issues.apache.org/jira/browse/CASSANDRA-2521?focusedCommentId=13050396&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13050396
>
>
>
> but my logs doesn't show the
I don't know if I understand correctly that UUIDs are good unless you have a
specific reading pattern. In the latter case you can develop a better
compound row key.
Yi
從我的 BlackBerry® 無線裝置
-Original Message-
From: Ramesh S
Date: Thu, 29 Sep 2011 16:26:05
To:
Reply-To: user@cassandr
Thanks Aaron.
Appreciate your valuable input/advice.
regards,
Ramesh
On Thu, Sep 29, 2011 at 4:21 PM, aaron morton wrote:
> UUID will be fine, LexicalUUID should be used for version 2,3,4 and 5
> UUID's. TimeUUID for version 1.
>
> A
>
> -
> Aaron Morton
> Freelance Cassandra Dev
UUID will be fine, LexicalUUID should be used for version 2,3,4 and 5 UUID's.
TimeUUID for version 1.
A
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 30/09/2011, at 5:48 AM, Ramesh S wrote:
> We have to assign Id for each item in our
As with any situation involving the un-dead, it really is the number of
Zombies, Mummies or Vampires that is the concern.
If you delete data there will always be tombstones. If you have a delete heavy
workload there will be more tombstones. This is why implementing a queue with
cassandra is a
If you are collecting time series data, and assuming the flying turtles we live
on that swim through time do not stop, you will want to partition your data.
(background http://www.slideshare.net/mattdennis/cassandra-data-modeling)
Lets say it makes sense for you to partition by month (may not be
I missed the subcomparator. Probably this is the solution.
create column family ProductCategory
with column_type = 'Super'
and comparator = UTF8Type
WITH subcomparator = UTF8Type
AND key_validation_class=UTF8Type
AND column_metadata = [
{column_name: subProdName, validation_class: UTF8Type
Thank you very much!
Just read some stuff in the wiki, such as limitations and secondary index.
Adding up to what you said, the search in large rows, by which I mean rows
with millions of columns, seams to be like searching normal hash instead of
btree style.
So model A it is!
Once again thank yo
I am trying to create a super column family using Cli command.
But I am not getting it.
The structure is
<>ProductCategory
<>#ProductType
<>#productCatId
+subProdName
+lenght
+width
I tried a lot many ways but I can't find the right way to get this done.
Something like this give me error - misma
So I need to read what I write before hitting send. Should have been,
"If A works for YOUR use case." and "Wide rows DON'T spread across nodes
well"
On 09/29/2011 02:34 PM, Jeremiah Jordan wrote:
If A works for our use case, it is a much better option. A given row
has to be read in full to r
If A works for our use case, it is a much better option. A given row
has to be read in full to return data from it, there used to be
limitations that a row had to fit in memory, but there is now code to
page through the data, so while that isn't a limitation any more, it
means rows that don't
I see now :
https://issues.apache.org/jira/browse/CASSANDRA-2521?focusedCommentId=13050396&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13050396
but my logs doesn't show the error from
MMappedSegmentedFile:
public static void initCleaner()
{
tr
I thought the need for GC was based on the code that sstables
deletions are tracked through weak/phantom references.
now I see that sstable is deleted in SStableReader.ReleaseReferences()
---> SegmentedFile.cleanup(),
if this is true, is GC still necessary?
Thanks
Yang
On Thu, Sep 29, 2011 at 11:
As the comment says,
// retry after GCing to force unmap of compacted SSTables
so they can be deleted
See also http://wiki.apache.org/cassandra/MemtableSSTable
Note that this has changed in 1.0:
https://issues.apache.org/jira/browse/CASSANDRA-2521
On Thu, Sep 29, 2011 at 1:14 PM, Yan
I personally have nothing but the highest praise for Elastic Search. I
haven't used either Solandra or Lucandra, however. But then why would I now
I use Elastic Search? AFAIK there's nothing that I would want to do with ES
that I can't. And it has really simple configuration and clustering that
jus
Table.getDataFileLocation() calls StorageService.requestGC(), which
calls System.gc()
I think I saw one GC in my log due to this code, and it caused a JVM
freeze for 14 seconds.
why do we need to call explicit GC ?
thanks
Yang
[Times: user=2.81 sys=0.00, real=0.46 secs]
3472.943: [GC 3472.943:
Yes, because the cache is not full so there is no pressure to remove
old entries.
On Thu, Sep 29, 2011 at 11:52 AM, Sanjeev Kulkarni
wrote:
> Hi Thamizh,
> Thanks for the answer.
> I understand the part about the Key cache capacity being 20 which is the
> default value.
> But Key cache size b
What would be the best approach
A) millions of ~2Kb rows, where each row could have ~6 columns
B) hundreds of ~100Gb rows, where each row could have ~1million columns
Considerarions:
Most entries will be searched for (read+write) at least once a day but no
more than 3 times a day.
Cheap hardware a
Hi Thamizh,
Thanks for the answer.
I understand the part about the Key cache capacity being 20 which is the
default value.
But Key cache size being 99k? Does this represent that cassandra has
allocated 99k for key cache even though the actual keys are far less?
On Thu, Sep 29, 2011 at 3:47 AM,
We have to assign Id for each item in our database. Item is linked to geo
location and hence would need hundreds of millions of Ids.
So is LexicalUUID a good option ?
regards,
Ramesh
From a usability standpoint, elastic search is looking promising. I'll have to
get more info through use on it's distribution as well.
Thanks
:)
On 28/09/2011, at 14:01 PM, Mohit Anchlia wrote:
> look at elasticsearch too. It shards differently.
>
> On Wed, Sep 28, 2011 at 1:45 PM, Rafael Al
Jonathan/Aaron,
Thank you guy's reply, I will change GCGracePeriod to 1 day to see what will
happen.
Is there a way to purge tombstones at anytime? because if tombstones affect
performance, we want them to be purged right away, not after GCGracePeriod.
We know all the nodes are up, and we can do
please check [default@unknown] help create column family;
These are default values, until you explicitly mentioned on CF creation.
Regards,
Thamizhannal
From: Sanjeev Kulkarni
To: user@cassandra.apache.org
Sent: Thursday, 29 September 2011 10:33 AM
Subject: n
If the retrieval of URL is based on "TimeUUID". Then Model C with
ByteOrderedPartitioner and rowkey as long type of "TimeUUID" can be correct
choice and it helps you to apply range query based on TimeUUID.
Regards,
Thamizhannal P
From: M Vieira
To: user@cassa
I am out of the office until 10/03/2011.
I will be out office on 09/29/2011 and 09/30/2011.
I will respond to your message when I return to the office
Note: This is an automated response to your message "Cassandra data
modeling" sent on 9/29/11 3:24:58.
This is the only notification you wi
Hi,
Thanks a lot Aaron. I had killed this process before it 8 attempts got
completed. Now it it uploading records.
Now, I am trying to perform parallel SSTable upload using 2 instances(127.0.0.2
& 127.0.0.3). Now the first instance is uploading and the second
instance(127.0.0.3), I am getting
I'm trying to get my head around Cassandra data modeling, but I can't quite
see what would be the best approach to the problem I have.
The supposed scenario:
You have around 100 domains, each domain have from few hundreds to millions
of possible URLs (think of different combinations of GET args,
ex
37 matches
Mail list logo