Skip to content
\n

For each tier, we are interested in the disk_used and samples

\n

For this agents restart frequency (at least daily) we get

\n\n

Note that the numbers for Tier 0 and 1 are \"better\" that what is listed in the documentation (Tier 2 is worse). This is the reason also for mentioning Usually On Disk

\n

For an agent with more tiers, a similar approach can be used for calculation

\n

Note: These disk_used numbers do not include agent metadata not related to data collection itself and is stored per metric (metric name, chart, context etc) but that doesn't impact higher tiers anyway

\n

I hope this helps

","upvoteCount":2,"url":"https://github.com/netdata/netdata/discussions/18616#discussioncomment-10760972"}}}
Discussion options

You must be logged in to vote

Hi @aldem

It has to do with update frequency of the higher tiers. Higher tiers are updated every X iterations of data points collected of the previous tiers.

So in the default configuration tier 0 collects metrics every second, tier 1 will store 1 point every 60 metrics of tier 0 etc.

As collected metrics fill data pages for each tier, the pages are grouped into "blocks", compressed and stored to disk

Agent behavior (for example frequent restarts) will result in data pages with just a few data points. This can affect compression, but it can also hurt the compression ratio (because even for 1 data point there is associated metadata that needs to be stored)

As an example for a running agent…

Replies: 2 comments

Comment options

You must be logged in to vote
0 replies
Answer selected by aldem
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants