However, if I were you I would avoid that... Maybe I will place a url to S3 or GFS in Cassandra
Best, Sergio On Tue, May 31, 2022, 4:10 PM Sergio <lapostadiser...@gmail.com> wrote: > You have to split it by yourself > Best, > Sergio > > On Tue, May 31, 2022, 3:56 PM Andria Trigeorgis <an.trigeo...@gmail.com> > wrote: > >> Thank you for your prompt reply! >> So, I have to split the blob into chunks by myself, or there is any >> fragmentation mechanism in Cassandra? >> >> >> On 31 May 2022, at 4:44 PM, Dor Laor <d...@scylladb.com> wrote: >> >> On Tue, May 31, 2022 at 4:40 PM Andria Trigeorgi <an.trigeo...@gmail.com> >> wrote: >> >>> Hi, >>> >>> I want to write large blobs in Cassandra. However, when I tried to write >>> more than a 256MB blob, I got the message: >>> "Error from server: code=2200 [Invalid query] message=\"Request is too >>> big: length 268435580 exceeds maximum allowed length 268435456.\"". >>> >>> I tried to change the variables "max_value_size_in_mb" and " >>> native_transport_max_frame_size_in_mb" of the file " >>> /etc/cassandra/cassandra.yaml" to 512, but I got a >>> ConnectionRefusedError error. What am I doing wrong? >>> >> >> You sent a large blob ;) >> >> This limitation exists to protect you as a user. >> The DB can store such blobs but it will incur a large and unexpected >> latency, not just >> for the query but also for under-the-hood operations, like backup and >> repair. >> >> Best is not to store such large blobs in Cassandra or chop them into >> smaller >> units, let's say 10MB pieces and re-assemble in the app. >> >> >>> >>> Thank you in advance, >>> >>> Andria >>> >> >>