Thanks Tyler for sharing this. It is exactly what I was looking for to know.

best,
/Shahab

On Thu, Oct 23, 2014 at 5:37 PM, Tyler Hobbs <ty...@datastax.com> wrote:

> CASSANDRA-8091 (Stress tool creates too large batches) is relevant:
> https://issues.apache.org/jira/browse/CASSANDRA-8091
>
> On Thu, Oct 23, 2014 at 6:28 AM, shahab <shahab.mok...@gmail.com> wrote:
>
>> OK, Thanks again Jens.
>>
>> best,
>> /Shahab
>>
>> On Thu, Oct 23, 2014 at 1:22 PM, Jens Rantil <jens.ran...@tink.se> wrote:
>>
>>> Hi again Shabab,
>>>
>>> Yes, it seems that way. I have no experience with the “cassandra stress
>>> tool”, but wouldn’t be surprised if the batch size could be tweaked.
>>>
>>> Cheers,
>>> Jens
>>>
>>> ——— Jens Rantil Backend engineer Tink AB Email: jens.ran...@tink.se
>>> Phone: +46 708 84 18 32 Web: www.tink.se Facebook Linkedin Twitter
>>>
>>>
>>> On Thu, Oct 23, 2014 at 10:00 AM, shahab <shahab.mok...@gmail.com>
>>> wrote:
>>>
>>>> Thanks Jens for the comments.
>>>>
>>>> As I am trying "cassandra stress tool", does it mean that the tool is
>>>> executing batch of "Insert" statements (probably hundreds, or thousands)
>>>>  to the cassandra (for the sake of stressing Cassnadra ?
>>>>
>>>> best,
>>>> /Shahab
>>>>
>>>> On Wed, Oct 22, 2014 at 8:14 PM, Jens Rantil <jens.ran...@tink.se>
>>>> wrote:
>>>>
>>>>>  Shabab,
>>>>>
>>>>> Apologize for the late answer.
>>>>>
>>>>> On Mon, Oct 6, 2014 at 2:38 PM, shahab <shahab.mok...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> But do you mean that inserting columns with large size (let's say a
>>>>>> text with 20-30 K) is potentially problematic in Cassandra?
>>>>>
>>>>>
>>>>> AFAIK, the size _warning_ you are getting relates to the size of the
>>>>> batch of prepared statements (INSERT INTO mykeyspace.mytable VALUES
>>>>> (?,?,?,?)). That is, it has nothing to do with the actual content of
>>>>> your row. 20-30 K shouldn't be a problem. But it's considered good 
>>>>> practise
>>>>> to split larger files (maybe > 5 MB into chunks) since it makes operations
>>>>> easier to your cluster more likely to spread more evenly across cluster.
>>>>>
>>>>>
>>>>>> What shall i do if I want columns with large size?
>>>>>
>>>>>
>>>>> Just don't insert to many rows in a single batch and you should be
>>>>> fine. Like Shane's JIRA ticket said, the warning is to let you know you 
>>>>> are
>>>>> not following best practice when adding too many rows in a single batch. 
>>>>> It
>>>>> can create bottlenecks in a single Cassandra node.
>>>>>
>>>>> Cheers,
>>>>> Jens
>>>>>
>>>>> --
>>>>> Jens Rantil
>>>>> Backend engineer
>>>>> Tink AB
>>>>>
>>>>> Email: jens.ran...@tink.se
>>>>> Phone: +46 708 84 18 32
>>>>> Web: www.tink.se
>>>>>
>>>>>  Facebook <https://www.facebook.com/#!/tink.se> Linkedin
>>>>> <http://www.linkedin.com/company/2735919?trk=vsrp_companies_res_photo&trkInfo=VSRPsearchId%3A1057023381369207406670%2CVSRPtargetId%3A2735919%2CVSRPcmpt%3Aprimary>
>>>>>  Twitter <https://twitter.com/tink>
>>>>>
>>>>
>>>>
>>>
>>
>
>
> --
> Tyler Hobbs
> DataStax <http://datastax.com/>
>

Reply via email to