Is the node selection based on key deterministic across multiple clients?
If it is, that sounds plausible. For this particular workload it's
definitely possible to have a hot key / spot, but it was surprising it
wasn't three nodes that got hot, it was just one.
On Mon, Nov 14, 2016 at 6:26 PM, Ale
Hi Abhishek,
Can you check if you're getting the same behavior on this cluster using
nodetool commands to start repair ? (don't forget to add --full in order to
make sure you're not running incremental repair, if that's indeed what
you're doing with reaper).
Could you also open an issue on github
Hi Abhishek
The article with the futex bug description lists the solution, which is to
upgrade to a version of RHEL or CentOS that have the specified patch.
What help do you specifically need? If you need help upgrading the OS I
would look at the documentation for RHEL or CentOS.
Ben
On Mon, 14
Hi,
We are seeing an issue where the system CPU is shooting off to a figure or
> 90% when the cluster is subjected to a relatively high write workload i.e
4k wreq/secs.
2016-11-14T13:27:47.900+0530 Process summary
process cpu=695.61%
application cpu=676.11% (*user=200.63% sys=475.49%) **<== V
Hi All,
we tried sequential repair on very small table having only 20 Rows using
the reaper tool. But the repair got stuck while generating the snapshot.
Same when we tried with Parallel repair then run was working fine in the
begining for few segments but later it got stuck in the compaction an
I'm wondering if what you are seeing is
https://datastax-oss.atlassian.net/browse/PYTHON-643 (that could still be a
sign of a potential data hotspot)
On Sun, Nov 13, 2016 at 10:57 PM, Andrew Bialecki <
andrew.biale...@klaviyo.com> wrote:
> We're using the "default" TokenAwarePolicy. Our nodes are
URI comes in pretty handy ;
video://videoprovider:codecSomething:myConverter:videoId
Or XRI but what Michael said.
> On Nov 14, 2016, at 11:59 AM, Michael Shuler wrote:
>
> Forward thinking, I would also suggest not storing the full URL, just
> the video ID of some sort. The application cod
Forward thinking, I would also suggest not storing the full URL, just
the video ID of some sort. The application code can create the URL as
needed, using the ID. If the full URL is stored in Cassandra and some
day in the future, the video file storage system needs to be changed,
this would require
For the record, there is an interesting use case of globo.com using
Cassandra to store video payload and stream live video at scale (in
particular, the FIFA World Cup + Olympics), but it's a pretty
non-conventional/advanced use case:
-
https://leandromoreira.com.br/2015/04/26/fifa-2014-world-cup-li
Some time ago, I stumbled across this:
https://github.com/chrislusf/seaweedfs
It is an open source implementation of Facebooks Haystack design. Have no
experience yet but we will evaluate it as a blob-store to replace our
Mogile-FS installation which stores over one billion images. From my point
of
I am truly sorry, Raghavendra. It didn't occur to me that you could be a
beginner.
On Mon, Nov 14, 2016 at 11:46 PM, Jon Haddad
wrote:
> Think about it like this. You just started using Cassandra for the first
> time. You have a question, you find there’s a mailing list, and you ask.
> You hav
While Cassandra *can* be used this way, I don’t recommend it. It’s going to be
far cheaper and easier to maintain to store data in an Object store like S3,
like Oskar recommended.
> On Nov 14, 2016, at 10:16 AM, l...@airstreamcomm.net wrote:
>
> We store videos and files in Cassandra by chunki
Think about it like this. You just started using Cassandra for the first time.
You have a question, you find there’s a mailing list, and you ask. You have
zero experience with the DB and are an outsider to a community. You ask
anyways, because it’s where the Apache website says to go. You g
Seconded. It is completely unhelpful to spam this list. Please stop.
> On Nov 14, 2016, at 12:21 PM, Jon Haddad wrote:
>
> You’ve asked a lot of questions on this mailing list, and you’ve gotten help
> on a ton of beginner issues. Making fun of someone for asking similar
> beginner question
You’ve asked a lot of questions on this mailing list, and you’ve gotten help on
a ton of beginner issues. Making fun of someone for asking similar beginner
questions is not cool at all. Cut it out.
> On Nov 14, 2016, at 10:13 AM, Ali Akhtar wrote:
>
> Another solution could be to print the
Excuse me? I did not make fun of anyone. I gave valid suggestions that are
all theoretically possible.
If it came off in a condescending way, i am genuinely sorry.
On 14 Nov 2016 11:22 pm, "Jon Haddad" wrote:
> You’ve asked a lot of questions on this mailing list, and you’ve gotten
> help on a
We store videos and files in Cassandra by chunking them into small portions and
saving them as blobs. As for video you could track the file byte offset of
each chunk and request the relevant pieces when scrubbing to a particular
portion of the video.
> On Nov 14, 2016, at 11:02 AM, raghavend
Another solution could be to print the raw bytes to paper, and write the
page numbers to cassandra. Playback will be challenging with this method
however, unless interns are available to transcribe the papers back to a
digital format.
On Mon, Nov 14, 2016 at 11:06 PM, Ali Akhtar wrote:
> The vid
The video can be written to floppy diskettes, and the serial numbers of the
diskettes can be written to cassandra.
On Mon, Nov 14, 2016 at 11:00 PM, Oskar Kjellin
wrote:
> The actual video is not stored in Cassandra. You need to use a proper
> origin like s3.
>
> Although you can probably store
The actual video is not stored in Cassandra. You need to use a proper origin
like s3.
Although you can probably store it in Cassandra, it's not a good idea.
Sent from my iPhone
> On 14 nov. 2016, at 18:02, raghavendra vutti
> wrote:
>
> Hi,
>
> Just wanted to know How does hulu or netfli
Hi,
Just wanted to know How does hulu or netflix store videos in cassandra.
Do they just use references to the video files in the form of URL's and
store in the DB??
could someone please me on this.
Thanks,
Raghavendra.
Hi Folks,
I have a table that has lot of tombstones generated and has caused inconsistent
data across various datacenters. we run anti-entropy repairs and also have
read_repair_chance tuned-up during our non busy hours. But yet when we try to
compare data residing in various replicas across DCs,
Hi Boying,
I agree with Vladimir.If compaction is not compacting the two sstables with
updates soon, disk space issues will be wasted. For example, if the updates are
not closer in time, first update might be in a big table by the time second
update is being written in a new small table. STCS wo
nodetool cfstats will show it per table.
The bloom filter / compression data is typically (unless you have very unusual
settings in your schema) 1-3GB each per TB of data, so with 235’ish GB/server,
it’s unlikely bloom filter or compression data.
The memTable is AT LEAST 1MB per columnfam
Hi Boying,
UPDATE write new value with new time stamp. Old value is not tombstone, but
remains until compaction. gc_grace_period is not related to this.
Best regards, Vladimir Yudovin,
Winguzone - Hosted Cloud Cassandra
Launch your cluster in minutes.
On Mon, 14 Nov 2016 03:02:21
Hi, All,
Will the Cassandra generates a new tombstone when updating a column by using
CQL update statement?
And is there any way to get the number of tombstones of a column family since
we want to void generating
too many tombstones within gc_grace_period?
Thanks
Boying
26 matches
Mail list logo