Specifying the retention period

Rahul Amaram rahul.amaram at vizury.com
Thu Sep 11 20:43:37 CEST 2014


I don't think so. Look at the table below (14.30 is the current time).

0: 13-14
1: 12-13
2: 11-12
3: 10-11
4: 9-10
5: 8-9
6: 7-8
7: 6-7
8: 5-6
9: 4-5
10: 3-4
11: 2-3
12: 1-2
13: 0-1
14: 23-0
15: 22-23
16: 21-22
17: 20-21
18: 19-20
19: 18-19
20: 17-18
21: 16-17
22: 15-16
23: 14-15

:)

- Rahul.

On Thursday 11 September 2014 05:19 PM, Anders Håål wrote:
> Index 0 is the last calculated hour, in your case  at time 2:30 index 
> 0 holds the avg for data between 1 and 2 a clock. This means, if my 
> calculation is correct, that the previous avg between 2 and 3 the day 
> before is at index 22.
> Anders
>
> On 09/11/2014 07:15 AM, Rahul Amaram wrote:
>> Also, let us say, that the current time is 2.30 and that I want the 
>> average of all the values between 2.00 and 3.00 the previous day, I'd 
>> probably have to use
>>
>> $$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[23]
>>
>> rather than
>>
>> $$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[24]
>>
>> Am I right ?
>>
>> Thanks,
>> Rahul.
>>
>> On Thursday 11 September 2014 10:39 AM, Rahul Amaram wrote:
>>> Ok. So would 
>>> $$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[24] refer to 
>>> the average of the all the values ONLY in the 24th hour before the 
>>> current time?
>>>
>>> On Thursday 11 September 2014 10:30 AM, Anders Håål wrote:
>>>> Hi Amaram,
>>>> I think you just need to remove the minus sign when using the 
>>>> aggregated. Minus is used for time, like back in time, and just a 
>>>> integer without minus and a time indicator is an index. Check out 
>>>> http://www.bischeck.org/wp-content/uploads/2014/06/Bischeck_configuration_guide.html#toc-Chapter-4. 
>>>>
>>>> You can also use redis-cli to explore the data in the cache. The 
>>>> key in the redis is the same as the service definition.
>>>> Anders
>>>>
>>>> On 09/11/2014 06:38 AM, Rahul Amaram wrote:
>>>>> Ok. I am facing another issue. I have been running bischeck with 
>>>>> the aggregate function for more than a day. I am using the below 
>>>>> threshold function.
>>>>>
>>>>> <threshold>avg($$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[-24],$$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[-168],$$HOSTNAME$$-$$SERVICENAME$$/H/avg-$$SERVICEITEMNAME$$[-336])</threshold> 
>>>>>
>>>>>
>>>>> and it doesn't seem to work. I am expecting that the first 
>>>>> aggregate value should be available.
>>>>>
>>>>> Instead if I use the below threshold function (I know this is not 
>>>>> related to aggregate)
>>>>>
>>>>> avg($$HOSTNAME$$-$$SERVICENAME$$-$$SERVICEITEMNAME$$[-24H],$$HOSTNAME$$-$$SERVICENAME$$-$$SERVICEITEMNAME$$[-168H],$$HOSTNAME$$-$$SERVICENAME$$-$$SERVICEITEMNAME$$[-336H]) 
>>>>>
>>>>>
>>>>> the threshold is calcuated fine, which is just the first value as 
>>>>> the remaining two values are not in cache.
>>>>>
>>>>> How can I debug why aggregate is not working?
>>>>>
>>>>> Thanks,
>>>>> Rahul.
>>>>>
>>>>> On Wednesday 10 September 2014 04:53 PM, Anders Håål wrote:
>>>>>> Thanks - got the ticket.
>>>>>> I will update progress on the bug ticket, but its good that the 
>>>>>> work around works.
>>>>>> Anders
>>>>>>
>>>>>> On 09/10/2014 01:20 PM, Rahul Amaram wrote:
>>>>>>> That indeed seems to be the problem. Using count rather than period
>>>>>>> seems to address the issue. Raised a ticket -
>>>>>>> http://gforge.ingby.com/gf/project/bischeck/tracker/?action=TrackerItemEdit&tracker_item_id=259 
>>>>>>>
>>>>>>> .
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Rahul.
>>>>>>>
>>>>>>> On Wednesday 10 September 2014 04:02 PM, Anders Håål wrote:
>>>>>>>> This looks like a bug. Could you please report it on
>>>>>>>> http://gforge.ingby.com/gf/project/bischeck/tracker/ in the Bugs
>>>>>>>> tracker. You need a account but its just a sign up and you get an
>>>>>>>> email confirmation.
>>>>>>>> Can you try to use maxcount for purging instead as a work 
>>>>>>>> around? Just
>>>>>>>> calculate your maxcount based on the scheduling interval you use.
>>>>>>>> Anders
>>>>>>>>
>>>>>>>> On 09/10/2014 12:17 PM, Rahul Amaram wrote:
>>>>>>>>> Following up on the earlier topic, I am seeing the below 
>>>>>>>>> errors related
>>>>>>>>> to cache purge. Any idea on what might be causing this? I 
>>>>>>>>> don't see any
>>>>>>>>> other errors in log related to metrics.
>>>>>>>>>
>>>>>>>>> 2014-09-10 12:12:00.001 ; INFO ; 
>>>>>>>>> DefaultQuartzScheduler_Worker-5 ;
>>>>>>>>> com.ingby.socbox.bischeck.configuration.CachePurgeJob ; 
>>>>>>>>> CachePurge
>>>>>>>>> purging 180
>>>>>>>>> 2014-09-10 12:12:00.003 ; INFO ; 
>>>>>>>>> DefaultQuartzScheduler_Worker-5 ;
>>>>>>>>> com.ingby.socbox.bischeck.configuration.CachePurgeJob ; 
>>>>>>>>> CachePurge
>>>>>>>>> executed in 1 ms
>>>>>>>>> 2014-09-10 12:12:00.003 ; ERROR ; 
>>>>>>>>> DefaultQuartzScheduler_Worker-5 ;
>>>>>>>>> org.quartz.core.JobRunShell ; Job DailyMaintenance.CachePurge 
>>>>>>>>> threw an
>>>>>>>>> unhandled Exception: java.lang.NullPointerException: null
>>>>>>>>>          at
>>>>>>>>> com.ingby.socbox.bischeck.cache.provider.redis.LastStatusCache.trim(LastStatusCache.java:1250) 
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>          at
>>>>>>>>> com.ingby.socbox.bischeck.configuration.CachePurgeJob.execute(CachePurgeJob.java:140) 
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2014-09-10 12:12:00.003 ; ERROR ; 
>>>>>>>>> DefaultQuartzScheduler_Worker-5 ;
>>>>>>>>> org.quartz.core.ErrorLogger ; Job (DailyMaintenance.CachePurge 
>>>>>>>>> threw an
>>>>>>>>> exception.org.quartz.SchedulerException: Job threw an unhandled
>>>>>>>>> exception.
>>>>>>>>>          at org.quartz.core.JobRunShell.run(JobRunShell.java:224)
>>>>>>>>>          at
>>>>>>>>> org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:557) 
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Caused by: java.lang.NullPointerException: null
>>>>>>>>>          at
>>>>>>>>> com.ingby.socbox.bischeck.cache.provider.redis.LastStatusCache.trim(LastStatusCache.java:1250) 
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>          at
>>>>>>>>> com.ingby.socbox.bischeck.configuration.CachePurgeJob.execute(CachePurgeJob.java:140) 
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Here is my cache configuration:
>>>>>>>>>
>>>>>>>>>      <cache>
>>>>>>>>>        <aggregate>
>>>>>>>>>          <method>avg</method>
>>>>>>>>>          <useweekend>true</useweekend>
>>>>>>>>>          <retention>
>>>>>>>>>            <period>H</period>
>>>>>>>>>            <offset>720</offset>
>>>>>>>>>          </retention>
>>>>>>>>>          <retention>
>>>>>>>>>            <period>D</period>
>>>>>>>>>            <offset>30</offset>
>>>>>>>>>          </retention>
>>>>>>>>>        </aggregate>
>>>>>>>>>
>>>>>>>>>        <purge>
>>>>>>>>>          <offset>30</offset>
>>>>>>>>>          <period>D</period>
>>>>>>>>>        </purge>
>>>>>>>>>      </cache>
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Rahul.
>>>>>>>>> On Monday 08 September 2014 08:39 PM, Anders Håål wrote:
>>>>>>>>>> Great if you can make a debian package, and I understand that 
>>>>>>>>>> you can
>>>>>>>>>> not commit. The best thing would be integrated to our build 
>>>>>>>>>> process
>>>>>>>>>> where we use ant.
>>>>>>>>>>
>>>>>>>>>> if the purging is based on time then it could happen that 
>>>>>>>>>> data is
>>>>>>>>>> removed from the cache since the logic is based on time 
>>>>>>>>>> relative to
>>>>>>>>>> now. To avoid it you should increase the purge time before 
>>>>>>>>>> you start
>>>>>>>>>> bischeck. And just a comment on your last sentence Redis TTl 
>>>>>>>>>> is never
>>>>>>>>>> used :)
>>>>>>>>>> Anders
>>>>>>>>>>
>>>>>>>>>> On 09/08/2014 02:09 PM, Rahul Amaram wrote:
>>>>>>>>>>> I would be more than happy to give you guys a testimonial. 
>>>>>>>>>>> However, we
>>>>>>>>>>> have just taken this live and would like to see its performance
>>>>>>>>>>> before I
>>>>>>>>>>> give a testimonial.
>>>>>>>>>>>
>>>>>>>>>>> Also, if time permits, I'll try to bundle this for Debian 
>>>>>>>>>>> (I'm a
>>>>>>>>>>> Debian
>>>>>>>>>>> maintainer). I can't commit on a timeline right away though :).
>>>>>>>>>>>
>>>>>>>>>>> Also, just to make things explicitly clear. I understand 
>>>>>>>>>>> that the
>>>>>>>>>>> below
>>>>>>>>>>> service item ttl has nothing to do with Redis TTL. But If I 
>>>>>>>>>>> stop my
>>>>>>>>>>> bischeck server for a day or two, then would any of my 
>>>>>>>>>>> metrics get
>>>>>>>>>>> lost?
>>>>>>>>>>> Or would I have to increase th Redis TTL for this.
>>>>>>>>>>>
>>>>>>>>>>> Regards,
>>>>>>>>>>> Rahul.
>>>>>>>>>>>
>>>>>>>>>>> On Monday 08 September 2014 04:09 PM, Anders Håål wrote:
>>>>>>>>>>>> Glad that it clarified how to configure the cache section. 
>>>>>>>>>>>> I will
>>>>>>>>>>>> make
>>>>>>>>>>>> a blog post on this in the mean time, until we have a updated
>>>>>>>>>>>> documentation. I agree with you that the structure of the
>>>>>>>>>>>> configuration is a bit "heavy", so ideas and input is 
>>>>>>>>>>>> appreciated.
>>>>>>>>>>>>
>>>>>>>>>>>> Regarding redis ttl, this is a redis feature we do not use. 
>>>>>>>>>>>> The ttl
>>>>>>>>>>>> mentioned in my mail is managed by bischeck. Redis ttl on 
>>>>>>>>>>>> linked list
>>>>>>>>>>>> do not work on individual nodes in a redis linked list.
>>>>>>>>>>>>
>>>>>>>>>>>> Currently the bischeck installer should work for ubuntu,
>>>>>>>>>>>> redhat/centos
>>>>>>>>>>>> and debian. There is currently no plans to make 
>>>>>>>>>>>> distribution packages
>>>>>>>>>>>> like rpm or deb. I know op5 (www.op5.com) that bundles 
>>>>>>>>>>>> Bischeck
>>>>>>>>>>>> make a
>>>>>>>>>>>> bischeck rpm. It would be super if there is any one that 
>>>>>>>>>>>> like to do
>>>>>>>>>>>> this for the project.
>>>>>>>>>>>> When it comes to packaging we have done a bit of work to 
>>>>>>>>>>>> create
>>>>>>>>>>>> docker
>>>>>>>>>>>> containers, but its still experimental.
>>>>>>>>>>>>
>>>>>>>>>>>> I also encourage you, if you think bischeck support your 
>>>>>>>>>>>> monitoring
>>>>>>>>>>>> effort, to write a small testimony that we can put on the 
>>>>>>>>>>>> site.
>>>>>>>>>>>> Regards
>>>>>>>>>>>> Anders
>>>>>>>>>>>>
>>>>>>>>>>>> On 09/08/2014 11:30 AM, Rahul Amaram wrote:
>>>>>>>>>>>>> Thanks Anders. This explains precisely why my data was 
>>>>>>>>>>>>> getting
>>>>>>>>>>>>> purged
>>>>>>>>>>>>> after 16 hours (30 values per hour * 1 hours = 480). It 
>>>>>>>>>>>>> would be
>>>>>>>>>>>>> great
>>>>>>>>>>>>> if you could update the documentation with this info. The 
>>>>>>>>>>>>> entire
>>>>>>>>>>>>> setup
>>>>>>>>>>>>> and configuration itself takes time to get a hold on and 
>>>>>>>>>>>>> detailed
>>>>>>>>>>>>> documentation would be very helpful.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Also, another quick question? Right now, I believe the 
>>>>>>>>>>>>> Redis TTL is
>>>>>>>>>>>>> set
>>>>>>>>>>>>> to 2000 seconds. Does this mean that if I don't receive 
>>>>>>>>>>>>> data for a
>>>>>>>>>>>>> particular serviceitem (or service or host) for a 2000 
>>>>>>>>>>>>> seconds, the
>>>>>>>>>>>>> data
>>>>>>>>>>>>> related to it is lost?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Also, any plans for bundling this with distributions such 
>>>>>>>>>>>>> as Debian?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>> Rahul.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Monday 08 September 2014 02:04 PM, Anders Håål wrote:
>>>>>>>>>>>>>> Hi Rahul,
>>>>>>>>>>>>>> Thanks for the question and feedback on the 
>>>>>>>>>>>>>> documentation. Great to
>>>>>>>>>>>>>> hear that you think Bischeck is awesome. If you do not
>>>>>>>>>>>>>> understand how
>>>>>>>>>>>>>> it works by reading the documentation you are probably not
>>>>>>>>>>>>>> alone, and
>>>>>>>>>>>>>> we should consider it a documentation bug.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> In 1.0.0 we introduce the concept that you asking about 
>>>>>>>>>>>>>> and it
>>>>>>>>>>>>>> really
>>>>>>>>>>>>>> two different independent features.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Lets start with cache purging.
>>>>>>>>>>>>>> Collected monitoring data, metrics, are kept in the cache 
>>>>>>>>>>>>>> (redis
>>>>>>>>>>>>>> from
>>>>>>>>>>>>>> 1.0.0) as a linked lists. There is one linked list per 
>>>>>>>>>>>>>> service
>>>>>>>>>>>>>> definition, like host1-service1-serviceitem1. Prior to 1.0.0
>>>>>>>>>>>>>> all the
>>>>>>>>>>>>>> linked lists had the same size that was defined with the 
>>>>>>>>>>>>>> property
>>>>>>>>>>>>>> lastStatusCacheSize. But in 1.0.0 we made that 
>>>>>>>>>>>>>> configurable so it
>>>>>>>>>>>>>> could be defined per service definition.
>>>>>>>>>>>>>> To enable individual cache configurations we added a 
>>>>>>>>>>>>>> section called
>>>>>>>>>>>>>> <cache> in the serviceitem section of the bischeck.xml. 
>>>>>>>>>>>>>> Like many
>>>>>>>>>>>>>> other configuration options in 1.0.0 the cache section could
>>>>>>>>>>>>>> have the
>>>>>>>>>>>>>> specific values or point to a template that could be shared.
>>>>>>>>>>>>>> To manage the size of the cache , or to be more specific 
>>>>>>>>>>>>>> the linked
>>>>>>>>>>>>>> list size, we defined the <purge> section. The purge 
>>>>>>>>>>>>>> section can
>>>>>>>>>>>>>> have
>>>>>>>>>>>>>> two different configurations. The first is defining the 
>>>>>>>>>>>>>> max size of
>>>>>>>>>>>>>> the cache linked list.
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <purge>
>>>>>>>>>>>>>>    <maxcount>1000</maxcount>
>>>>>>>>>>>>>>   </purge>
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The second options is to define the “time to live” for the
>>>>>>>>>>>>>> metrics in
>>>>>>>>>>>>>> the cache.
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <purge>
>>>>>>>>>>>>>>    <offset>10</offset>
>>>>>>>>>>>>>>    <period>D</period>
>>>>>>>>>>>>>>   </purge>
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>> In the above example we set the time to live to 10 days. 
>>>>>>>>>>>>>> So any
>>>>>>>>>>>>>> metrics older then this period will be removed. The 
>>>>>>>>>>>>>> period can have
>>>>>>>>>>>>>> the following values:
>>>>>>>>>>>>>> H - hours
>>>>>>>>>>>>>> D - days
>>>>>>>>>>>>>> W - weeks
>>>>>>>>>>>>>> Y - year
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The two option are mutual exclusive. You have to chose 
>>>>>>>>>>>>>> one for each
>>>>>>>>>>>>>> serviceitem or cache template.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> If no cache directive is define for a serviceitem the 
>>>>>>>>>>>>>> property
>>>>>>>>>>>>>> lastStatusCacheSize will be used. It's default value is 500.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hopefully this explains the cache purging.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The next question was related to aggregations which has 
>>>>>>>>>>>>>> nothing
>>>>>>>>>>>>>> to do
>>>>>>>>>>>>>> with purging, but it's configured in the same <cache> 
>>>>>>>>>>>>>> section. The
>>>>>>>>>>>>>> idea with aggregations was to create an automatic way to 
>>>>>>>>>>>>>> aggregate
>>>>>>>>>>>>>> metrics on the level of an hour, day, week and month. The
>>>>>>>>>>>>>> aggregation
>>>>>>>>>>>>>> functions current supported is average, max and min.
>>>>>>>>>>>>>> Lets say you have a service definition of the format
>>>>>>>>>>>>>> host1-service1-serviceitem1. When you enable an average 
>>>>>>>>>>>>>> (avg)
>>>>>>>>>>>>>> aggregation you will automatically get the following new 
>>>>>>>>>>>>>> service
>>>>>>>>>>>>>> definitions
>>>>>>>>>>>>>> host1-service1/H/avg-serviceitem1
>>>>>>>>>>>>>> host1-service1/D/avg-serviceitem1
>>>>>>>>>>>>>> host1-service1/W/avg-serviceitem1
>>>>>>>>>>>>>> host1-service1/M/avg-serviceitem1
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The configuration you need to achive the above average
>>>>>>>>>>>>>> aggregations is:
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>avg</method>
>>>>>>>>>>>>>>   </aggregate>
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> If you like to combine it with the above descibed purging 
>>>>>>>>>>>>>> your
>>>>>>>>>>>>>> configuration would look like:
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>avg</method>
>>>>>>>>>>>>>>   </aggregate>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>   <purge>
>>>>>>>>>>>>>>    <offset>10</offset>
>>>>>>>>>>>>>>    <period>D</period>
>>>>>>>>>>>>>>   </purge>
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The new aggregated service definitions,
>>>>>>>>>>>>>> host1-service1/H/avg-serviceitem1, etc, will have their 
>>>>>>>>>>>>>> own cache
>>>>>>>>>>>>>> entries and can be used in threshold configurations and 
>>>>>>>>>>>>>> virtual
>>>>>>>>>>>>>> services like any other service definitions. For example 
>>>>>>>>>>>>>> in a
>>>>>>>>>>>>>> threshold hours section we could define
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> <hours hoursID="2">
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>   <hourinterval>
>>>>>>>>>>>>>>     <from>09:00</from>
>>>>>>>>>>>>>>     <to>12:00</to>
>>>>>>>>>>>>>> <threshold>host1-service1/H/avg-serviceitem1[0]*0.8</threshold> 
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>   </hourinterval>
>>>>>>>>>>>>>>   ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> This would mean that we use the average value for
>>>>>>>>>>>>>> host1-service1-serviceitem1  for the period of the last 
>>>>>>>>>>>>>> hour.
>>>>>>>>>>>>>> Aggregations are calculated hourly, daily, weekly and 
>>>>>>>>>>>>>> monthly.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> By default weekends metrics are not included in the 
>>>>>>>>>>>>>> aggrgation
>>>>>>>>>>>>>> calculation. This can be enabled by setting the
>>>>>>>>>>>>>> <useweekend>true</useweekend>:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>avg</method>
>>>>>>>>>>>>>> <useweekend>true</useweekend>
>>>>>>>>>>>>>>   </aggregate>
>>>>>>>>>>>>>>   ….
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> This will create aggregated service definitions with the 
>>>>>>>>>>>>>> following
>>>>>>>>>>>>>> name standard:
>>>>>>>>>>>>>> host1-service1/H/avg/weekend-serviceitem1
>>>>>>>>>>>>>> host1-service1/D/avg/weekend-serviceitem1
>>>>>>>>>>>>>> host1-service1/W/avg/weekend-serviceitem1
>>>>>>>>>>>>>> host1-service1/M/avg/weekend-serviceitem1
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> You can also have multiple entries like:
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>avg</method>
>>>>>>>>>>>>>> <useweekend>true</useweekend>
>>>>>>>>>>>>>>   </aggregate>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>max</method>
>>>>>>>>>>>>>>   </aggregate>
>>>>>>>>>>>>>>   ….
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> So how long time will the aggregated values be kept in the
>>>>>>>>>>>>>> cache? By
>>>>>>>>>>>>>> default we save
>>>>>>>>>>>>>> Hour aggregation for 25 hours
>>>>>>>>>>>>>> Daily aggregations for 7 days
>>>>>>>>>>>>>> Weekly aggregations for 5 weeks
>>>>>>>>>>>>>> Monthly aggregations for 1 month
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> These values can be override but they can not be lower 
>>>>>>>>>>>>>> then the
>>>>>>>>>>>>>> default. Below you have an example where we save the 
>>>>>>>>>>>>>> aggregation
>>>>>>>>>>>>>> for
>>>>>>>>>>>>>> 168 hours, 60 days and 53 weeks.
>>>>>>>>>>>>>> <cache>
>>>>>>>>>>>>>>   <aggregate>
>>>>>>>>>>>>>>     <method>avg</method>
>>>>>>>>>>>>>> <useweekend>true</useweekend>
>>>>>>>>>>>>>>     <retention>
>>>>>>>>>>>>>>       <period>H</period>
>>>>>>>>>>>>>>       <offset>168</offset>
>>>>>>>>>>>>>>     </retention>
>>>>>>>>>>>>>>     <retention>
>>>>>>>>>>>>>>      <period>D</period>
>>>>>>>>>>>>>>       <offset>60</offset>
>>>>>>>>>>>>>>     </retention>
>>>>>>>>>>>>>>     <retention>
>>>>>>>>>>>>>>       <period>W</period>
>>>>>>>>>>>>>>       <offset>53</offset>
>>>>>>>>>>>>>>     </retention>
>>>>>>>>>>>>>> </aggregate>
>>>>>>>>>>>>>>   ….
>>>>>>>>>>>>>> </cache>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I hope this makes it a bit less confusing. What is clear 
>>>>>>>>>>>>>> to me is
>>>>>>>>>>>>>> that
>>>>>>>>>>>>>> we need to improve the documentation in this area.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Looking forward to your feedback.
>>>>>>>>>>>>>> Anders
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 09/08/2014 06:02 AM, Rahul Amaram wrote:
>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>> I am trying to setup the bischeck plugin for our 
>>>>>>>>>>>>>>> organization. I
>>>>>>>>>>>>>>> have
>>>>>>>>>>>>>>> configured most part of it except for the cache 
>>>>>>>>>>>>>>> retention period.
>>>>>>>>>>>>>>> Here
>>>>>>>>>>>>>>> is what I want - I want to store every value which has been
>>>>>>>>>>>>>>> generated
>>>>>>>>>>>>>>> during the past 1 month. The reason being my threshold is
>>>>>>>>>>>>>>> currently
>>>>>>>>>>>>>>> calculated as the average of the metric value during the 
>>>>>>>>>>>>>>> past 4
>>>>>>>>>>>>>>> weeks at
>>>>>>>>>>>>>>> the same time of the day.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> So, how do I define the cache template for this? If I don't
>>>>>>>>>>>>>>> define any
>>>>>>>>>>>>>>> cache template, for how many days is the data kept?
>>>>>>>>>>>>>>> Also, how does the aggregrate function work and and what 
>>>>>>>>>>>>>>> does the
>>>>>>>>>>>>>>> purge
>>>>>>>>>>>>>>> Maxitems signify?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I've gone through the documentation but it wasn't clear. 
>>>>>>>>>>>>>>> Looking
>>>>>>>>>>>>>>> forward
>>>>>>>>>>>>>>> to a response.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Bischeck is one awesome plugin. Keep up the great work.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>>> Rahul.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>>
>
>


-- 



More information about the Bischeck-users mailing list