There is a fair amount of overhead in the Thrift structures for columns and
mutations, so that's a pretty large mutation.

In general, you'll see better performance inserting many small batch
mutations in parallel.

On Fri, Aug 10, 2012 at 2:04 AM, Jin Lei <jehovah.l...@gmail.com> wrote:

> Sorry, something is wrong with my previous problem description. The fact
> is that the cassandra deny my requests when I try to insert 50k rows
> (rather than 50k columns) into a column family at one time. Each row with 1
> column.
>
> 2012/8/10 Jin Lei <jehovah.l...@gmail.com>
>
>> Hello everyone,
>> I'm a novice to cassandra and meet a problem recently.
>> I want to insert over 50k columns into cassandra at one time, total size
>> of which doesn't exceed 16MB, but the database return an exception as
>> follows.
>>
>> [E 120809 15:37:31 service:1251] error in write to database
>>     Traceback (most recent call last):
>>       File "/home/stoneiii/mycode/src/user/service.py", line 1248, in
>> flush_mutator
>>         self.mutator.send()
>>
>>       File "/home/stoneiii/mycode/pylib/pycassa/batch.py", line 127, in
>> send
>>
>>         conn.batch_mutate(mutations, write_consistency_level)
>>       File "/home/stoneiii/gaia2/pylib/pycassa/pool.py", line 145, in
>> new_f
>>         return new_f(self, *args, **kwargs)
>>       File "/home/stoneiii/mycode/pylib/pycassa/pool.py", line 145, in
>> new_f
>>         return new_f(self, *args, **kwargs)
>>       File "/home/stoneiii/mycode/pylib/pycassa/pool.py", line 145, in
>> new_f
>>         return new_f(self, *args, **kwargs)
>>       File "/home/stoneiii/mycode/pylib/pycassa/pool.py", line 145, in
>> new_f
>>         return new_f(self, *args, **kwargs)
>>       File "/home/stoneiii/mycode/pylib/pycassa/pool.py", line 145, in
>> new_f
>>         return new_f(self, *args, **kwargs)
>>       File "/home/stoneiii/mycode/pylib/pycassa/pool.py", line 140, in
>> new_f
>>         (self._retry_count, exc.__class__.__name__, exc))
>>     MaximumRetryException: Retried 6 times. Last failure was error:
>> [Errno 104] Connection reset by peer
>>
>> Since cassandra supports 2 billion of columns in one table, why can't I
>> insert 50k columns in this way? Or what settings should I adjust to break
>> this limit?
>> Thanks for any hint in advance!
>>
>>
>>
>>
>


-- 
Tyler Hobbs
DataStax <http://datastax.com/>

Reply via email to