Yes, sorry about that. Wednesday morning 9am PT On Tue, Apr 16, 2019 at 3:26 AM Benedict Elliott Smith <bened...@apache.org> wrote:
> Just to confirm, this is on Wednesday? > > > On 15 Apr 2019, at 22:38, Jon Haddad <j...@jonhaddad.com> wrote: > > > > Hey all, > > > > I've set up a Zoom call for 9AM Pacific time. Everyone's welcome to > join. > > > > https://zoom.us/j/189920888 > > > > Looking forward to a good discussion on how we can all pitch in on > > getting 4.0 out the door. > > > > Jon > > > > On Sat, Apr 13, 2019 at 9:14 AM Jonathan Koppenhofer > > <j...@koppedomain.com> wrote: > >> > >> Wednesday would work for me. > >> > >> We use and (slightly) contribute to tlp tools. We are platform testing > and > >> beginning 4.0 testing ourselves, so an in person overview would be > great! > >> > >> On Sat, Apr 13, 2019, 8:48 AM Aleksey Yeshchenko > <alek...@apple.com.invalid> > >> wrote: > >> > >>> Wednesday and Thursday, either, at 9 AM pacific WFM. > >>> > >>>> On 13 Apr 2019, at 13:31, Stefan Miklosovic < > >>> stefan.mikloso...@instaclustr.com> wrote: > >>>> > >>>> Hi Jon, > >>>> > >>>> I would like be on that call too but I am off on Thursday. > >>>> > >>>> I am from Australia so 5pm London time is ours 2am next day so your > >>>> Wednesday morning is my Thursday night. Wednesday early morning so > >>>> your Tuesday morning and London's afternoon would be the best. > >>>> > >>>> Recording the thing would be definitely helpful too. > >>>> > >>>> On Sat, 13 Apr 2019 at 07:45, Jon Haddad <j...@jonhaddad.com> wrote: > >>>>> > >>>>> I'd be more than happy to hop on a call next week to give you both > >>>>> (and anyone else interested) a tour of our dev tools. Maybe > something > >>>>> early morning on my end, which should be your evening, could work? > >>>>> > >>>>> I can set up a Zoom conference to get everyone acquainted. We can > >>>>> record and post it for any who can't make it. > >>>>> > >>>>> I'm thinking Tuesday, Wednesday, or Thursday morning, 9AM Pacific > (5pm > >>>>> London)? If anyone's interested please reply with what dates work. > >>>>> I'll be sure to post the details back here with the zoom link in case > >>>>> anyone wants to join that didn't get a chance to reply, as well as a > >>>>> link to the recorded call. > >>>>> > >>>>> Jon > >>>>> > >>>>> On Fri, Apr 12, 2019 at 10:41 AM Benedict Elliott Smith > >>>>> <bened...@apache.org> wrote: > >>>>>> > >>>>>> +1 > >>>>>> > >>>>>> I’m also just as excited to see some standardised workloads and test > >>> bed. At the moment we’re benefiting from some large contributors doing > >>> their own proprietary performance testing, which is super valuable and > >>> something we’ve lacked before. But I’m also keen to see some more > >>> representative workloads that are reproducible by anybody in the > community > >>> take shape. > >>>>>> > >>>>>> > >>>>>>> On 12 Apr 2019, at 18:09, Aleksey Yeshchenko > >>> <alek...@apple.com.INVALID> wrote: > >>>>>>> > >>>>>>> Hey Jon, > >>>>>>> > >>>>>>> This sounds exciting and pretty useful, thanks. > >>>>>>> > >>>>>>> Looking forward to using tlp-stress for validating 15066 > performance. > >>>>>>> > >>>>>>> We should touch base some time next week to pick a comprehensive > set > >>> of workloads and versions, perhaps? > >>>>>>> > >>>>>>> > >>>>>>>> On 12 Apr 2019, at 16:34, Jon Haddad <j...@jonhaddad.com> wrote: > >>>>>>>> > >>>>>>>> I don't want to derail the discussion about Stabilizing Internode > >>>>>>>> Messaging, so I'm starting this as a separate thread. There was a > >>>>>>>> comment that Josh made [1] about doing performance testing with > real > >>>>>>>> clusters as well as a lot of microbenchmarks, and I'm 100% in > support > >>>>>>>> of this. We've been working on some tooling at TLP for the last > >>>>>>>> several months to make this a lot easier. One of the goals has > been > >>>>>>>> to help improve the 4.0 testing process. > >>>>>>>> > >>>>>>>> The first tool we have is tlp-stress [2]. It's designed with a > "get > >>>>>>>> started in 5 minutes" mindset. My goal was to ship a stress tool > >>> that > >>>>>>>> ships with real workloads out of the box that can be easily > tweaked, > >>>>>>>> similar to how fio allows you to design a disk workload and tweak > it > >>>>>>>> with paramaters. Included are stress workloads that stress LWTs > (two > >>>>>>>> different types), materialized views, counters, time series, and > >>>>>>>> key-value workloads. Each workload can be modified easily to > change > >>>>>>>> compaction strategies, concurrent operations, number of > partitions. > >>>>>>>> We can run workloads for a set number of iterations or a custom > >>>>>>>> duration. We've used this *extensively* at TLP to help our > customers > >>>>>>>> and most of our blog posts that discuss performance use it as > well. > >>>>>>>> It exports data to both a CSV format and auto sets up prometheus > for > >>>>>>>> metrics collection / aggregation. As an example, we were able to > >>>>>>>> determine that the compression length set on the paxos tables > imposes > >>>>>>>> a significant overhead when using the Locking LWT workload, which > >>>>>>>> simulates locking and unlocking of rows. See CASSANDRA-15080 for > >>>>>>>> details. > >>>>>>>> > >>>>>>>> We have documentation [3] on the TLP website. > >>>>>>>> > >>>>>>>> The second tool we've been working on is tlp-cluster [4]. This > tool > >>>>>>>> is designed to help provision AWS instances for the purposes of > >>>>>>>> testing. To be clear, I don't expect, or want, this tool to be > used > >>>>>>>> for production environments. It's designed to assist with the > >>>>>>>> Cassandra build process by generating deb packages or re-using the > >>>>>>>> ones that have already been uploaded. Here's a short list of the > >>>>>>>> things you'll care about: > >>>>>>>> > >>>>>>>> 1. Create instances in AWS for Cassandra using any instance size > and > >>>>>>>> number of nodes. Also create tlp-stress instances and a box for > >>>>>>>> monitoring > >>>>>>>> 2. Use any available build of Cassandra, with a quick option to > >>> change > >>>>>>>> YAML config. For example: tlp-stress use 3.11.4 -c > >>>>>>>> concurrent_writes:256 > >>>>>>>> 3. Do custom builds just by pointing to a local Cassandra git > repo. > >>>>>>>> They can be used the same way as #2. > >>>>>>>> 4. tlp-stress is automatically installed on the stress box. > >>>>>>>> 5. Everything's installed with pure bash. I considered something > >>> more > >>>>>>>> complex, but since this is for development only, it turns out the > >>>>>>>> simplest tool possible works well and it means it's easily > >>>>>>>> configurable. Just drop in your own bash script starting with a > >>>>>>>> number in a XX_script_name.sh format and it gets run. > >>>>>>>> 6. The monitoring box is running Prometheus. It auto scrapes > >>>>>>>> Cassandra using the Instaclustr metrics library. > >>>>>>>> 7. Grafana is also installed automatically. There's a couple > sample > >>>>>>>> graphs there now. We plan on having better default graphs soon. > >>>>>>>> > >>>>>>>> For the moment it installs java 8 only but that should be easily > >>>>>>>> fixable to use java 11 to test ZGC (it's on my radar). > >>>>>>>> > >>>>>>>> Documentation for tlp-cluster is here [5]. > >>>>>>>> > >>>>>>>> There's still some things to work out in the tool, and we've been > >>>>>>>> working hard to smooth out the rough edges. I still haven't > >>> announced > >>>>>>>> anything WRT tlp-cluster on the TLP blog, because I don't think > it's > >>>>>>>> quite ready for public consumption, but I think the folks on this > >>> list > >>>>>>>> are smart enough to see the value in it even if it has a few warts > >>>>>>>> still. > >>>>>>>> > >>>>>>>> I don't consider myself familiar enough with the networking patch > to > >>>>>>>> give it a full review, but I am qualified to build tools to help > test > >>>>>>>> it and go through the testing process myself. From what I can > tell > >>>>>>>> the patch is moving the codebase in a positive direction and I'd > like > >>>>>>>> to help build confidence in it so we can get it merged in. > >>>>>>>> > >>>>>>>> We'll continue to build out and improve the tooling with the goal > of > >>>>>>>> making it easier for people to jump into the QA side of things. > >>>>>>>> > >>>>>>>> Jon > >>>>>>>> > >>>>>>>> [1] > >>> > https://lists.apache.org/thread.html/742009c8a77999f4b62062509f087b670275f827d0c1895bf839eece@%3Cdev.cassandra.apache.org%3E > >>>>>>>> [2] https://github.com/thelastpickle/tlp-stress > >>>>>>>> [3] http://thelastpickle.com/tlp-stress/ > >>>>>>>> [4] https://github.com/thelastpickle/tlp-cluster > >>>>>>>> [5] http://thelastpickle.com/tlp-cluster/ > >>>>>>>> > >>>>>>>> > --------------------------------------------------------------------- > >>>>>>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > >>>>>>>> For additional commands, e-mail: dev-h...@cassandra.apache.org > >>>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > --------------------------------------------------------------------- > >>>>>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > >>>>>>> For additional commands, e-mail: dev-h...@cassandra.apache.org > >>>>>>> > >>>>>> > >>>>> > >>>>> --------------------------------------------------------------------- > >>>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > >>>>> For additional commands, e-mail: dev-h...@cassandra.apache.org > >>>>> > >>>> > >>>> --------------------------------------------------------------------- > >>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > >>>> For additional commands, e-mail: dev-h...@cassandra.apache.org > >>>> > >>> > >>> > >>> --------------------------------------------------------------------- > >>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > >>> For additional commands, e-mail: dev-h...@cassandra.apache.org > >>> > >>> > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > > For additional commands, e-mail: dev-h...@cassandra.apache.org > > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org > For additional commands, e-mail: dev-h...@cassandra.apache.org > >