Thanks for your answer but I am just curious about: i)How do you identify the size of the object which you are going to chunk?
ii) While reading or updating how it is going to read all those chunks? Vikas On Thu, Oct 20, 2016 at 9:25 PM, Justin Cameron <jus...@instaclustr.com> wrote: > You can, but it is not really very efficient or cost-effective. You may > encounter issues with streaming, repairs and compaction if you have very > large blobs (100MB+), so try to keep them under 10MB if possible. > > I'd suggest storing blobs in something like Amazon S3 and keeping just the > bucket name & blob id in Cassandra. > > On Thu, 20 Oct 2016 at 12:03 Vikas Jaiman <er.vikasjai...@gmail.com> > wrote: > >> Hi, >> >> Normally people would like to store smaller values in Cassandra. Is there >> anyone using it to store for larger values (e.g 500KB or more) and if so >> what are the issues you are facing . I Would like to know the tweaks also >> which you are considering. >> >> Thanks, >> Vikas >> > -- > > Justin Cameron > > Senior Software Engineer | Instaclustr > > > > > This email has been sent on behalf of Instaclustr Pty Ltd (Australia) and > Instaclustr Inc (USA). > > This email and any attachments may contain confidential and legally > privileged information. If you are not the intended recipient, do not copy > or disclose its content, but please reply to this email immediately and > highlight the error to the sender and then immediately delete the message. > > --