This the multi-page printable view of this section. Click here to print.

Return to the regular view of this page.

Observability Platform

Observability Platform

We are the team responsible (mainly) for:

1 - Observability Platform Team SRE Processes

Observability Platform Team SRE Processes

This document explains a few processes related to operating our production software together with AppSRE team. All operations can be summarised as the Site Reliability Engineering part of our work.

Goals

The main goals of our SRE work within this team is to deploy and operate software that meets functional (API and performance (Telemeter SLO, MST SLO) requirements of our customers.

Currently, we offer internal service called RHOBS that is used across the company.

Releasing Changes / Patches

In order to maintain our software on production reliably we need to be able to test, release and deploy our changes rapidly.

We can divide things we change into a few categories. Let’s elaborate all processes per each category:

Software Libraries

Library is a project that is not deployed directly, but rather it is a dependency for micro-service application we deploy. For example https://github.com/prometheus/client_golang or https://github.com/grpc-ecosystem/go-grpc-middleware

Testing:

  • Linters, unit and integration tests on every PR.

Releasing

  • GitHub release using git tag (RC first, then major/minor or patch release).

Deploying

Software for (Micro) Services

Software for (micro) services usually lives in many separate open source repositories in GitHub e.g https://github.com/thanos-io/thanos, https://github.com/observatorium/api.

Testing:

  • Linters, unit and integration tests on every PR.

Releasing

  • GitHub release using git tag (RC first, then major/minor or patch release).
  • Building docker images in quay.io and dockerhub (backup) using CI for each tag or main branch PR.

Deploying

Configuration

All or configuration is rooted in https://github.com/rhobs/configuration configuration templates written in jsonnet. It can be then overridden by defined parameters in app-interface saas files .

Testing:

  • Building jsonnet resources, linting jsonnet, validating openshift templates.
  • Validation on App-interface side against Kubernetes YAMLs compatibility

Releasing

  • Merge to main and get the commit hash you want to deploy.

Deploying

  • Make sure you are on RH VPN.
  • Propose PR to https://gitlab.cee.redhat.com/service/app-interface with the change of ref for the desired environment to desired commit sha is app-interface Saas file for desired tenant (telemeter or MST) or change environment parameter.
  • Ask team for review. If change is impacting production heavily, notify AppSRE.
    • If only saas.yaml file was changed /lgtm from Observability Platform team is enough for PR to get merged automatically.
    • If any other file was changed, AppSRE engineer has to lgtm it.
  • When merged, CI will deploy the changes to namespace specified in saas.yaml e.g to production.

NOTE: Don’t change both production and staging in the same MR. NOTE: Deploy to production only changed that were previously in staging (automation for this TBD).

You can see the version change:

Monitoring Resources

Grafana Dashboards are defined here: https://github.com/rhobs/configuration/tree/main/observability/dashboards Alerts and Rules here: https://github.com/rhobs/configuration/blob/main/observability/prometheusrules.jsonnet

Testing:

  • ):

Releasing

  • Merge to main and get the commit hash you want to deploy.

Deploying

On-Call Rotation

Currently, RHOBS services are supported 24h/7d by two teams:

  • AppSRE team for infrastructure and “generalist” support. They are our first incident response point. They try to support our stack as far as runbooks and general knowledge allows.
  • Observability Platform Team (Dev on-call) for incidents impacting SLA outside of the AppSRE expertise (bug fixes, more complex troubleshooting). We are notified when AppSRE needs.

Incident Handling

This is the process we as the Observability Team try to follow during incident response.

The incident occurs when any of our services violates SLO we set with our stakeholders. Refer to Telemeter SLO and MST SLO for details on SLA.

NOTE: Following procedure applies to both Production and Staging. Many teams e.g SubWatch depends on working staging, so follow similar process as production. The only difference is that we do not need to mitigate of fix staging issues outside of office hours.

Potential Trigger Points:

  • You got notified on Slack by AppSRE team or paged through Pager Duty.
  • You got notified about potential SLA violation by the customer: Unexpected responses, things that worked before do not work now etc.
  • You touch production for unrelated reasons and noticed something worrying (error logs, un-monitored resource etc).
  1. If you are not on-call notify Observability Platform on-call engineer. If you are on-call, on-call engineer is not present or you agreed that you will handle this incident, go to step 2.

  2. Straight away, create JIRA for potential incident. Don’t think twice, it’s easy to create and essential to track incident later on. Fill the following parts:

    • Title: Symptom you see.

    • Type: Bug

    • Priority: Try to assess how important it is. If impacting production it’s a “Blocker”

    • Component: RHOBS

    • (Important) Label: incident no-qe

    • Description: Mention how you were notified (ideally with link to alert/Slack thread). Mention what you know so far.

      jira See example incident tickets here

  3. If AppSRE is not yet aware, drop link to created incident ticket to #sd-app-sre channel and notify @app-sre-primary and @observatorium-oncall. Don’t repeat yourself, ask everyone to follow the ticket comments.

    • AppSRE may or may not create dedicated channel, do communication efforts and start on-call Zoom meeting. We as the dev team don’t need to worry about those elements, go to step 4.
  4. Investigate possible mitigation. Ensure the problem is mitigated before focusing on root cause analysis.

    • Important: Note all performed actions and observations on created JIRA ticket via comments. This allows anyone to follow up what was checked and how. It is also essential for detailed Post Mortem / RCA process later on.
    • Note on JIRA ticket all automation or monitoring gaps you wished you had. This will be useful for actions after the incident.
  5. Ensure if incident is mitigated with AppSRE. Investigate root cause. If mitigation is applied and root cause is known, claim the incident being over.

  6. Note the potential long term fixes and ideas. Close the incident JIRA ticket.

After Incident

  1. After some time (within week), start a Post-mortem (RCA) document in Google Docs. Use following Google Docs RCA template. Put it in our RCA Team Google directory here. Link it in JIRA ticket too.
  2. Collaborate on Post Mortem. Make sure it is blameless but accurate. Share it as soon as possible with the team and AppSRE.
  3. Once done, schedule meeting with the Observability Platform and optionally AppSRE to discuss RCA/Post Mortem action items and effect.

Idea: If you have time, before sharing Post Mortem / RCA perform “Wheel of Misfortune”. Select an on-call engineer who was not participaing in incident and simulate the error by triggering root-cause in safe environment. Then meet together with team and allow engineer to coordinate simulated incident. Help on the way to share knowledge and insights. This is the best way to on-board people to production topics.

NOTE: Freely available SRE book is a good source of general patterns around in efficient incident management. Recommended read!