Thank you for your prompt reply! So, I have to split the blob into chunks by myself, or there is any fragmentation mechanism in Cassandra?
> On 31 May 2022, at 4:44 PM, Dor Laor <d...@scylladb.com> wrote: > > On Tue, May 31, 2022 at 4:40 PM Andria Trigeorgi <an.trigeo...@gmail.com > <mailto:an.trigeo...@gmail.com>> wrote: > Hi, > > I want to write large blobs in Cassandra. However, when I tried to write more > than a 256MB blob, I got the message: > "Error from server: code=2200 [Invalid query] message=\"Request is too big: > length 268435580 exceeds maximum allowed length 268435456.\"". > > I tried to change the variables "max_value_size_in_mb" and > "native_transport_max_frame_size_in_mb" of the file > "/etc/cassandra/cassandra.yaml" to 512, but I got a ConnectionRefusedError > error. What am I doing wrong? > > You sent a large blob ;) > > This limitation exists to protect you as a user. > The DB can store such blobs but it will incur a large and unexpected latency, > not just > for the query but also for under-the-hood operations, like backup and repair. > > Best is not to store such large blobs in Cassandra or chop them into smaller > units, let's say 10MB pieces and re-assemble in the app. > > > Thank you in advance, > > Andria