On Thu, Jun 15, 2023 at 9:01 AM Boris Behrens <[email protected]> wrote:
> Hi Ben, > I've read the link you gave me and I will try to answer the best I can. > > > What problem are you solving for? > I want to keep historical data for longer periods with less disk space > used. > I currently keep around 4 month of data for around 1.3T of disk space. > and I would love to have 1-2 years of data without having to buy 20TB > disks :) > 2 years of data in your case is only 8T. That's not a large amount of disk. > > > Why do you think you need this? > For debugging and monitoring I only need the last couple of days with > a raw resolution. And seeing trends can be downsampled (is this the > correct term?) because it is just historical trend data. > I think it's rather a use case than an actual problem I am working on. > I would like to see in the past how things changed and overlay it with > timestamps from changes. Things begin to looks way different when you > check the past. > Yes, downsampling is the correct term. But downsampling is not required to do this. Grafana supports special `$__interval` and `$__rate_interval` variables that can be used to automatically scale graphs based on desired time ranges. ` > > You say "a lot of endpoints", but what is a lot? 1,000? 10,000? 100,000? > 1,000,000? How many series? > A couple hundred endpoints with around 1.5m of series. But I don't > think that this information is helpful in the current context. > It is, because it gives us a sense of scale for your problem. You're running a small deployment, so no fancy changes are really necessary. For comparison, my $dayjob has about 900TiB of TSDB storage (Thanos object storage). Managing this is a different scale to a single node deployment. I think you're falling into the premature optimization trap. You could use a full Thanos deployment, as it supports exactly the kind of variable retention downsampling you want. But it's going to take a lot more work than just buying slightly bigger disks to store Prometheus on. > > Thank you for your time and effort to help me with my stupid little > problem. > Cheers > Boris > > Am Do., 15. Juni 2023 um 06:24 Uhr schrieb Ben Kochie <[email protected]>: > > > > My first question is "why?" (https://xyproblem.info/) > > > > What problem are you solving for? Why do you think you need this? > > > > You say "a lot of endpoints", but what is a lot? 1,000? 10,000? 100,000? > 1,000,000? How many series? > > > > On Thu, Jun 15, 2023 at 6:21 AM Boris <[email protected]> wrote: > >> > >> Dear prometheus community, > >> > >> my prometheus scrapes a lot of endpoints and I only need to 10s for a > month. After that I would like to rollup the data to 5m averages and after > 6 months I would like to have only hourly averages. > >> > >> So I can have quite a long time of data and can see changes over a long > time with the same dashboard. > >> > >> I am sorry to ask this question but it seems like I don't have the > correct search term. Downsampling seems to keep the raw data and all > searches lead me to timescaledb. > >> > >> Cheers > >> Boris > >> > >> -- > >> You received this message because you are subscribed to the Google > Groups "Prometheus Users" group. > >> To unsubscribe from this group and stop receiving emails from it, send > an email to [email protected]. > >> To view this discussion on the web visit > https://groups.google.com/d/msgid/prometheus-users/8D01579C-5696-40BB-9841-A1E3E0094DD1%40kervyn.de > . > > > > -- > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend > im groüen Saal. > -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/CABbyFmojP%2BsB4%2BdwGV%3DnJkSp-zVa%3Dn%2BXz5Qd1i3eOgAy%3DU_tsw%40mail.gmail.com.

