It's being lead by StreamNative, which provides a hosted Pulsar service. They haven't said it, but my gut's telling me that the end goal is to consolidate their serverless/hosted solution into a single cluster instead of needing to shard into multiple clusters.
- Nathan On Sun, Mar 16, 2025, at 19:31, Matt Pavlovich wrote: > I’d love to see the bill for storing metrics and logs for 1mm topics or > hundreds of millions of topics. > > That sounds more like a splash blog post vs a practical solution. > >> On Mar 8, 2025, at 1:06 PM, Nathan Clayton <apache....@nathanclayton.com> >> wrote: >> >> I'm not sure about Kafka, but Apache Pulsar is currently built to support up >> to 1mm topics. I know that there's active work on using Oxia instead of >> Zookeeper for the metadata store with the goal to support hundreds of >> millions of topics. >> >> That said, millions of topics in Artemis may be a bit of an anti-pattern. >> Depending on your use case, it may be better to use filter expressions on >> fewer queues [1]. >> >> Nathan >> >> [1] >> https://activemq.apache.org/components/artemis/documentation/latest/filter-expressions.html >> >> >> On Wed, Mar 5, 2025, at 12:45, William Crowell wrote: >>> I am wondering if Apache Kafka might be more feasible for something like >>> this. >>> >>> Regards, >>> >>> William Crowell >>> >>> From: Justin Bertram <jbert...@apache.org> >>> Date: Wednesday, March 5, 2025 at 3:26 PM >>> To: users@activemq.apache.org <users@activemq.apache.org> >>> Subject: Re: Maximum Amount of Topic/Queues Within Apache Artemis >>> That's it. >>> >>> >>> Justin >>> >>> On Wed, Mar 5, 2025 at 2:19 PM William Crowell >>> <wcrow...@perforce.com.invalid> wrote: >>> >>>> Justin, >>>> >>>> Never, I think I found it: >>>> >>>> “…the web console's behavior is configurable. Go to the "Preferences" >>>> (available from the menu in the top right) and click the "Jolokia" tab. >>>> Here you can turn off auto-refresh (i.e. "Update rate"). You can also >>>> decrease the amount of data fetched by lowering the "Max depth" and "Max >>>> collection size." >>>> >>>> Regards, >>>> >>>> William Crowell >>>> >>>> From: William Crowell <wcrow...@perforce.com.INVALID> >>>> Date: Wednesday, March 5, 2025 at 3:12 PM >>>> To: users@activemq.apache.org <users@activemq.apache.org> >>>> Subject: Re: Maximum Amount of Topic/Queues Within Apache Artemis >>>> Justin, >>>> >>>> Appreciate your reply. Your insight is greatly appreciated and invaluable. >>>> >>>> If you can provide more details on this, then I would appreciate it. I >>>> can search the archives for it as well. I am guessing there is some >>>> command line equivalent to represent the tree. >>>> >>>> What we have is several devices that have JMS clients that communicate >>>> with the broker on topics and durable queues. By “several devices” I mean >>>> maybe 100s or 1000s. >>>> >>>> Regards, >>>> >>>> William Crowell >>>> >>>> From: Justin Bertram <jbert...@apache.org> >>>> Date: Wednesday, March 5, 2025 at 2:57 PM >>>> To: users@activemq.apache.org <users@activemq.apache.org> >>>> Subject: Re: Maximum Amount of Topic/Queues Within Apache Artemis >>>> The broker itself doesn't impose any arbitrary limits on the number of >>>> addresses & queues. That said, there certainly are limits, the size of your >>>> heap probably being the most important. Every address and queue carry with >>>> it some memory overhead including the objects themselves but also related >>>> objects like JMX MBeans which enable management. >>>> >>>> The current management console will struggle by default with a huge number >>>> of addresses and/or queues due to the way the "tree" view works (i.e. >>>> refreshing the whole view on a regular basis). This is the main reason we >>>> removed this view in the new console [1]. That said, the console can be >>>> configured in such a way as to mitigate some of these problems. I can >>>> provide more details on that if necessary although it's been covered on >>>> this list a few times already. >>>> >>>> Regarding scalability...the broker was written with scalability in mind, >>>> but everything breaks down at some point. I'm interested to hear about your >>>> experiences if you go down this route especially regarding bottlenecks that >>>> you may find for your specific use-case. Keep us in the loop! >>>> >>>> >>>> Justin >>>> >>>> [1] >>>> https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-5319&data=05%7C02%7CWCrowell%40perforce.com%7C320052ba6ccd49ef708808dd5c240792%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638768032021655752%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=fiCybGio5R84gDL5NiT3uHHTrz%2FtKGU5k3AappaW%2FMc%3D&reserved=0<https://issues.apache.org/jira/browse/ARTEMIS-5319> >>>> < >>>> https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-5319&data=05%7C02%7CWCrowell%40perforce.com%7C320052ba6ccd49ef708808dd5c240792%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638768032021670531%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=x99X9Q4l3k5WDqM0eAAWMXIvYPR9ReXIlsBXyz3KBKo%3D&reserved=0<https://issues.apache.org/jira/browse/ARTEMIS-5319> >>>>> <https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-5319&data=05%7C02%7CWCrowell%40perforce.com%7C320052ba6ccd49ef708808dd5c240792%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638768032021679546%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=dtoeTZ5yS%2BtWShgjBs6svkrqwbvLSy6Y4OPAjTTsUwQ%3D&reserved=0<https://issues.apache.org/jira/browse/ARTEMIS-5319>> >>>> >>>> On Wed, Mar 5, 2025 at 12:33 PM William Crowell >>>> <wcrow...@perforce.com.invalid> wrote: >>>> >>>>> Good afternoon, >>>>> >>>>> Are there any limitations or concerns with creating millions of topics >>>> and >>>>> queues within Apache Artemis that have low volumes of messages in each >>>>> topic and queue? I do not think there are as I believe Artemis should be >>>>> able to handle this use case. Important question is: Is it scalable? >>>>> >>>>> Regards and have a great day, >>>>> >>>>> William Crowell >>>>> >>>>> >>>>> This e-mail may contain information that is privileged or confidential. >>>> If >>>>> you are not the intended recipient, please delete the e-mail and any >>>>> attachments and notify us immediately. >>>>> >>>>> >>>> >>>> >>>> CAUTION: This email originated from outside of the organization. Do not >>>> click on links or open attachments unless you recognize the sender and know >>>> the content is safe. >>>> >>>> >>>> This e-mail may contain information that is privileged or confidential. If >>>> you are not the intended recipient, please delete the e-mail and any >>>> attachments and notify us immediately. >>>> >>>> >>>> This e-mail may contain information that is privileged or confidential. If >>>> you are not the intended recipient, please delete the e-mail and any >>>> attachments and notify us immediately. >>>> >>>> >>> >>> >>> CAUTION: This email originated from outside of the organization. Do not >>> click on links or open attachments unless you recognize the sender and >>> know the content is safe. >>> >>> >>> This e-mail may contain information that is privileged or confidential. >>> If you are not the intended recipient, please delete the e-mail and any >>> attachments and notify us immediately. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: users-unsubscr...@activemq.apache.org >> For additional commands, e-mail: users-h...@activemq.apache.org >> For further information, visit: https://activemq.apache.org/contact >> >> > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: users-unsubscr...@activemq.apache.org > For additional commands, e-mail: users-h...@activemq.apache.org > For further information, visit: https://activemq.apache.org/contact --------------------------------------------------------------------- To unsubscribe, e-mail: users-unsubscr...@activemq.apache.org For additional commands, e-mail: users-h...@activemq.apache.org For further information, visit: https://activemq.apache.org/contact