Sentinel release 4.17 adds the ability to capture performance metrics for each monitor/test, in the Sentinel database.
Sentinel Usage Metrics
A new TestPerformanceMetrics table, which captures performance metrics for each monitor/test, has been added to the Sentinel database.
Two new configuration settings have been added to the Sentinel Configuration file: which need to be set:
- MonitorPerformanceMetricsEnabled: This needs to be set to True for the metrics to be collected.
- MonitorPerformanceMetricsPurgeTimeoutHours: Specifies how often to delete all the metrics from the table.
The table contains the following columns, which can be used to analyse performance issues:
COLUMN | DESCRIPTION |
TestId | The unique identifier of the test being analysed. |
TestVersionId | The version of the test. |
MonitorVersionId | The version of the monitor. |
TestRunTime | Time, in milliseconds, taken to run the test. |
DataFetchTime | Time, in milliseconds, taken to fetch the data. |
EventWriteTime | Time, in milliseconds, taken to an event to be written to the database. |
TotalRunTime | Total time, in milliseconds, taken for the monitor to run. |
FetchSourceTime | Time, in milliseconds, for the data source to respond to the fetch request. |
PreConditionRunTime | Time, in milliseconds, taken to for the pre-condition to run. |
EntityCount | The number of entities being processed. |
TriggerTime | The timestamp that the test was triggered. |
ProcessingStartTime | The timestamp that the test started being processed. |