gitlab-org--gitlab-foss/doc/development/snowplow/infrastructure.md

5.7 KiB

stage group info
Analytics Product Intelligence To determine the technical writer assigned to the Stage/Group associated with this page, see https://about.gitlab.com/handbook/engineering/ux/technical-writing/#assignments

Snowplow infrastructure

Snowplow events on GitLab SaaS fired by a tracker go through an AWS pipeline, managed by GitLab.

Event flow in the AWS pipeline

Every event goes through a collector, enricher, and pseudonymization lambda. The event is then dumped to S3 storage where it can be picked up by the Snowflake data warehouse.

Deploying and managing the infrastructure is automated using Terraform in the current Terraform repository.

graph LR
    GL[GitLab.com]-->COL

    subgraph aws-cloud[AWS]
    COL[Collector]-->|snowplow-raw-good|ENR
    COL[Collector]-->|snowplow-raw-bad|FRBE
        subgraph firehoserbe[Firehose]
        FRBE[AWS Lambda]
      end
    FRBE-->S3RBE

    ENR[Enricher]-->|snowplow-enriched-bad|FEBE
      subgraph firehoseebe[Firehose]
        FEBE[AWS Lambda]
      end
    FEBE-->S3EBE

    ENR[Enricher]-->|snowplow-enriched-good|FRGE
      subgraph firehosege[Firehose]
        FRGE[AWS Lambda]
      end
    FRGE-->S3GE
  end

    subgraph snowflake[Data warehouse]
    S3RBE[S3 raw-bad]-->BE[gitlab_bad_events]
    S3EBE[S3 enriched-bad]-->BE[gitlab_bad_events]
    S3GE[S3 output]-->GE[gitlab_events]
  end

See Snowplow technology 101 for Snowplow's own documentation and an overview how collectors and enrichers work.

Pseudonymization

In contrast to a typical Snowplow pipeline, after enrichment, GitLab Snowplow events go through a pseudonymization service in the form of an AWS Lambda service before they are stored in S3 storage.

Why events need to be pseudonymized

GitLab is bound by its obligations to community and by legal regulations to protect the privacy of its users.

GitLab must provide valuable insights for business decisions, and there is a need for a better understanding of different users' behavior patterns. The pseudonymization process helps you find a compromise between these two requirements.

Pseudonymization processes personally identifiable information inside a Snowplow event in an irreversible fashion maintaining deterministic output for given input, while masking any relation to that input.

How events are pseudonymized

Pseudonymization uses an allowlist that provides privacy by default. Therefore, each attribute received as part of a Snowplow event is pseudonymized unless the attribute is an allowed exception.

Pseudonymization is done using the HMAC-SHA256 keyed hash algorithm. Attributes are combined with a secret salt to replace each identifiable information with a pseudonym.

S3 bucket data lake to Snowflake

See Data team's Snowplow Overview for further details how data is ingested into our Snowflake data warehouse.

Monitoring

There are several tools that monitor Snowplow events tracking in different stages of the processing pipeline:

  • Product Intelligence Grafana dashboard monitors backend events sent from a GitLab.com instance to a collectors fleet. This dashboard provides information about:
    • The number of events that successfully reach Snowplow collectors.
    • The number of events that failed to reach Snowplow collectors.
    • The number of backend events that were sent.
  • AWS CloudWatch dashboard monitors the state of the events in a processing pipeline. The pipeline starts from Snowplow collectors, goes through to enrichers and pseudonymization, and then up to persistence in an S3 bucket. From S3, the events are imported into the Snowflake Data Warehouse. You must have AWS access rights to view this dashboard. For more information, see monitoring in the Snowplow Events pseudonymization service documentation.
  • Sisense dashboard provides information about the number of good and bad events imported into the Data Warehouse, in addition to the total number of imported Snowplow events.

For more information, see this video walk-through.