Thanks Tyler for sharing this. It is exactly what I was looking for to know.
best, /Shahab On Thu, Oct 23, 2014 at 5:37 PM, Tyler Hobbs <ty...@datastax.com> wrote: > CASSANDRA-8091 (Stress tool creates too large batches) is relevant: > https://issues.apache.org/jira/browse/CASSANDRA-8091 > > On Thu, Oct 23, 2014 at 6:28 AM, shahab <shahab.mok...@gmail.com> wrote: > >> OK, Thanks again Jens. >> >> best, >> /Shahab >> >> On Thu, Oct 23, 2014 at 1:22 PM, Jens Rantil <jens.ran...@tink.se> wrote: >> >>> Hi again Shabab, >>> >>> Yes, it seems that way. I have no experience with the “cassandra stress >>> tool”, but wouldn’t be surprised if the batch size could be tweaked. >>> >>> Cheers, >>> Jens >>> >>> ——— Jens Rantil Backend engineer Tink AB Email: jens.ran...@tink.se >>> Phone: +46 708 84 18 32 Web: www.tink.se Facebook Linkedin Twitter >>> >>> >>> On Thu, Oct 23, 2014 at 10:00 AM, shahab <shahab.mok...@gmail.com> >>> wrote: >>> >>>> Thanks Jens for the comments. >>>> >>>> As I am trying "cassandra stress tool", does it mean that the tool is >>>> executing batch of "Insert" statements (probably hundreds, or thousands) >>>> to the cassandra (for the sake of stressing Cassnadra ? >>>> >>>> best, >>>> /Shahab >>>> >>>> On Wed, Oct 22, 2014 at 8:14 PM, Jens Rantil <jens.ran...@tink.se> >>>> wrote: >>>> >>>>> Shabab, >>>>> >>>>> Apologize for the late answer. >>>>> >>>>> On Mon, Oct 6, 2014 at 2:38 PM, shahab <shahab.mok...@gmail.com> >>>>> wrote: >>>>> >>>>>> But do you mean that inserting columns with large size (let's say a >>>>>> text with 20-30 K) is potentially problematic in Cassandra? >>>>> >>>>> >>>>> AFAIK, the size _warning_ you are getting relates to the size of the >>>>> batch of prepared statements (INSERT INTO mykeyspace.mytable VALUES >>>>> (?,?,?,?)). That is, it has nothing to do with the actual content of >>>>> your row. 20-30 K shouldn't be a problem. But it's considered good >>>>> practise >>>>> to split larger files (maybe > 5 MB into chunks) since it makes operations >>>>> easier to your cluster more likely to spread more evenly across cluster. >>>>> >>>>> >>>>>> What shall i do if I want columns with large size? >>>>> >>>>> >>>>> Just don't insert to many rows in a single batch and you should be >>>>> fine. Like Shane's JIRA ticket said, the warning is to let you know you >>>>> are >>>>> not following best practice when adding too many rows in a single batch. >>>>> It >>>>> can create bottlenecks in a single Cassandra node. >>>>> >>>>> Cheers, >>>>> Jens >>>>> >>>>> -- >>>>> Jens Rantil >>>>> Backend engineer >>>>> Tink AB >>>>> >>>>> Email: jens.ran...@tink.se >>>>> Phone: +46 708 84 18 32 >>>>> Web: www.tink.se >>>>> >>>>> Facebook <https://www.facebook.com/#!/tink.se> Linkedin >>>>> <http://www.linkedin.com/company/2735919?trk=vsrp_companies_res_photo&trkInfo=VSRPsearchId%3A1057023381369207406670%2CVSRPtargetId%3A2735919%2CVSRPcmpt%3Aprimary> >>>>> Twitter <https://twitter.com/tink> >>>>> >>>> >>>> >>> >> > > > -- > Tyler Hobbs > DataStax <http://datastax.com/> >