- Existing SLIs
- Defining a new SLI
- Tracking operations for an SLI
- Using the SLI in service monitoring and alerts
GitLab Application Service Level Indicators (SLIs)
It is possible to define Service Level Indicators(SLIs) directly in the Ruby codebase. This keeps the definition of operations and their success close to the implementation and allows the people building features to easily define how these features should be monitored.
Existing SLIs
-
rails_request
-
global_search_apdex
-
global_search_error_rate
-
global_search_indexing_apdex
-
sidekiq_execution
Defining a new SLI
An SLI can be defined with the Gitlab::Metrics::Sli::Apdex
or
Gitlab::Metrics::Sli::ErrorRate
class. When you define an SLI, two
Prometheus counters
are emitted from the Rails application. Both counters work in broadly the same way and contain a total operation count. Apdex
uses a success rate to calculate a success ratio, and ErrorRate
uses an error rate to calculate an error ratio.
The following metrics are defined:
-
Gitlab::Metrics::Sli::Apdex.new('foo')
defines:-
gitlab_sli_foo_apdex_total
for the total number of measurements. -
gitlab_sli_foo_apdex_success_total
for the number of successful measurements.
-
-
Gitlab::Metrics::Sli::ErrorRate.new('foo')
defines:-
gitlab_sli_foo_total
for the total number of measurements. -
gitlab_sli_foo_error_total
for the number of error measurements. Because this metric is an error rate, errors are divided by the total number.
-
As shown in this example, they can share a base name (foo
in this example). We
recommend this when they refer to the same operation.
You should use Apdex
to measure the performance of successful operations. You don’t have to measure the performance of a failing request because that performance should be tracked with ErrorRate
. For example, you can measure whether a request is performing within a specified latency threshold.
You should use ErrorRate
to measure the rate of unsuccessful operations. For example, you can measure whether a failed request returns an HTTP status greater than or equal to 500
.
Before the first scrape, it is important to have initialized the SLI with all possible label-combinations. This avoid confusing results when using these counters in calculations.
To initialize an SLI, use the .initialize_sli
class method, for
example:
Gitlab::Metrics::Sli::Apdex.initialize_sli(:received_email, [
{
feature_category: :team_planning,
email_type: :create_issue
},
{
feature_category: :service_desk,
email_type: :service_desk
},
{
feature_category: :code_review_workflow,
email_type: :create_merge_request
}
])
Metrics must be initialized before they get scraped for the first time.
This currently happens during the on_master_start
lifecycle event.
Since this delays application readiness until metrics initialization returns, make sure the overhead
this adds is understood and acceptable.
Tracking operations for an SLI
Tracking an operation in the newly defined SLI can be done like this:
Gitlab::Metrics::Sli::Apdex[:received_email].increment(
labels: {
feature_category: :service_desk,
email_type: :service_desk
},
success: issue_created?
)
Calling #increment
on this SLI will increment the total Prometheus counter
gitlab_sli:received_email_apdex:total{ feature_category='service_desk', email_type='service_desk' }
If the success:
argument passed is truthy, then the success counter will also
be incremented:
gitlab_sli:received_email_apdex:success_total{ feature_category='service_desk', email_type='service_desk' }
For error rate SLIs, the equivalent argument is called error:
:
Gitlab::Metrics::Sli::ErrorRate[:merge].increment(
labels: {
merge_type: :fast_forward
},
error: !merge_success?
)
Using the SLI in service monitoring and alerts
When the application is emitting metrics for a new SLI, they need to be consumed from the metrics catalog to result in alerts, and included in the error budget for stage groups and GitLab.com’s overall availability.
Start by adding the new SLI to the Application-SLI library. After that, add the following information:
-
name
: the name of the SLI as defined in code. For examplereceived_email
. -
significantLabels
: an array of Prometheus labels that belong to the metrics. For example:["email_type"]
. If the significant labels for the SLI includefeature_category
, the metrics will also feed into the error budgets for stage groups. -
featureCategory
: if the SLI applies to a single feature category, you can specify it statically through this field to feed the SLI into the error budgets for stage groups. -
description
: a Markdown string explaining the SLI. It will be shown on dashboards and alerts. -
kind
: the kind of indicator. For examplesliDefinition.apdexKind
.
When done, run make generate
to generate recording rules for
the new SLI. This command creates recordings for all services
emitting these metrics aggregated over significantLabels
.
Open up a merge request with these changes and request review from a Scalability team member.
When these changes are merged, and the aggregations in Mimir recorded, query Mimir to see the success ratio of the new aggregated metrics. For example:
sum by (environment, stage, type)(application_sli_aggregation:rails_request:apdex:success:rate_1h)
/
sum by (environment, stage, type)(application_sli_aggregation:rails_request:apdex:weight:score_1h)
This shows the success ratio, which can guide you to set an appropriate SLO when adding this SLI to a service.
Then, add the SLI to the appropriate service
catalog file. For example, the web
service:
rails_requests:
sliLibrary.get('rails_request_apdex')
.generateServiceLevelIndicator({ job: 'gitlab-rails' })
To pass extra selectors and override properties of the SLI, see the service monitoring documentation.
SLIs with statically defined feature categories can already receive
alerts about the SLI in specified Slack channels. For more information, read the
alert routing documentation.
In this project
we are extending this so alerts for SLIs with a feature_category
label in the source metrics can also be routed.
For any question, don’t hesitate to create an issue in the Scalability issue tracker or come find us in #g_scalability on Slack.