Correct the spacing of unordered markdown lists in docs, to maintain standards of documentation.
14 KiB
Review Apps
Review Apps are automatically deployed by each pipeline, both in CE and EE.
How does it work?
CI/CD architecture diagram
Show mermaid source
graph TD build-qa-image -.->|once the `prepare` stage is done| gitlab:assets:compile review-build-cng -->|triggers a CNG-mirror pipeline and wait for it to be done| CNG-mirror review-build-cng -.->|once the `test` stage is done| review-deploy review-deploy -.->|once the `review` stage is done| review-qa-smokesubgraph 1. gitlab-ce/ee
prepare
stage build-qa-image endsubgraph 2. gitlab-ce/ee
test
stage gitlab:assets:compile -->|plays dependent job once done| review-build-cng endsubgraph 3. gitlab-ce/ee
review
stage review-deploy["review-deploy
Helm deploys the Review App using the Cloud
Native images built by the CNG-mirror pipeline.
Cloud Native images are deployed to thereview-apps-ce
orreview-apps-ee
Kubernetes (GKE) cluster, in the GCPgitlab-review-apps
project."] endsubgraph 4. gitlab-ce/ee
qa
stage review-qa-smoke[review-qa-smoke
gitlab-qa runs the smoke suite against the Review App.] endsubgraph CNG-mirror pipeline CNG-mirror>Cloud Native images are built]; end
Detailed explanation
- On every pipeline during the
test
stage, thegitlab:assets:compile
job is automatically started.- Once it's done, it starts the
review-build-cng
manual job since theCNG-mirror
pipeline triggered in the following step depends on it.
- Once it's done, it starts the
- The
review-build-cng
job triggers a pipeline in theCNG-mirror
project.- The
CNG-mirror
pipeline creates the Docker images of each component (e.g.gitlab-rails-ee
,gitlab-shell
,gitaly
etc.) based on the commit from the GitLab pipeline and stores them in its registry. - We use the
CNG-mirror
project so that theCNG
, (Cloud Native GitLab), project's registry is not overloaded with a lot of transient Docker images. - Note that the official CNG images are built by the
cloud-native-image
job, which runs only for tags, and triggers itself aCNG
pipeline.
- The
- Once the
test
stage is done, thereview-deploy
job deploys the Review App using the official GitLab Helm chart to thereview-apps-ce
/review-apps-ee
Kubernetes cluster on GCP.- The actual scripts used to deploy the Review App can be found at
scripts/review_apps/review-apps.sh
. - These scripts are basically
our official Auto DevOps scripts where the
default CNG images are overridden with the images built and stored in the
CNG-mirror
project's registry. - Since we're using the official GitLab Helm chart, this means you get a dedicated environment for your branch that's very close to what it would look in production.
- The actual scripts used to deploy the Review App can be found at
- Once the
review-deploy
job succeeds, you should be able to use your Review App thanks to the direct link to it from the MR widget. To log into the Review App, see "Log into my Review App?" below.
Additional notes:
- If the
review-deploy
job keep failing (note that we already retry it twice), please post a message in the#quality
channel and/or create a ~Quality ~bug issue with a link to your merge request. Note that the deployment failure can reveal an actual problem introduced in your merge request (i.e. this isn't necessarily a transient failure)! - If the
review-qa-smoke
job keep failing (note that we already retry it twice), please check the job's logs: you could discover an actual problem introduced in your merge request. You can also download the artifacts to see screenshots of the page at the time the failures occurred. If you don't find the cause of the failure or if it seems unrelated to your change, please post a message in the#quality
channel and/or create a ~Quality ~bug issue with a link to your merge request. - The manual
review-stop
in thetest
stage can be used to stop a Review App manually, and is also started by GitLab once a merge request's branch is deleted after being merged. - Review Apps are cleaned up regularly via a pipeline schedule that runs
the
schedule:review-cleanup
job. - The Kubernetes cluster is connected to the
gitlab-{ce,ee}
projects using GitLab's Kubernetes integration. This basically allows to have a link to the Review App directly from the merge request widget.
QA runs
On every pipeline in the qa
stage (which comes after the
review
stage), the review-qa-smoke
job is automatically started and it runs
the QA smoke suite.
You can also manually start the review-qa-all
: it runs the full QA suite.
Performance Metrics
On every pipeline in the qa
stage, the
review-performance
job is automatically started: this job does basic
browser performance testing using a
Sitespeed.io Container.
How to:
Log into my Review App
The default username is root
and its password can be found in the 1Password
secure note named gitlab-{ce,ee} Review App's root password.
Enable a feature flag for my Review App
- Open your Review App and log in as documented above.
- Create a personal access token.
- Enable the feature flag using the Feature flag API.
Find my Review App slug
- Open the
review-deploy
job. - Look for
Checking for previous deployment of review-*
. - For instance for
Checking for previous deployment of review-qa-raise-e-12chm0
, your Review App slug would bereview-qa-raise-e-12chm0
in this case.
Run a Rails console
- Filter Workloads by your Review App slug,
e.g.
review-qa-raise-e-12chm0
. - Find and open the
task-runner
Deployment, e.g.review-qa-raise-e-12chm0-task-runner
. - Click on the Pod in the "Managed pods" section, e.g.
review-qa-raise-e-12chm0-task-runner-d5455cc8-2lsvz
. - Click on the
KUBECTL
dropdown, thenExec
->task-runner
. - Replace
-c task-runner -- ls
with-it -- gitlab-rails console
from the default command or- Run
kubectl exec --namespace review-apps-ce review-qa-raise-e-12chm0-task-runner-d5455cc8-2lsvz -it -- gitlab-rails console
and- Replace
review-apps-ce
withreview-apps-ee
if the Review App is running EE, and - Replace
review-qa-raise-e-12chm0-task-runner-d5455cc8-2lsvz
with your Pod's name.
- Replace
- Run
Dig into a Pod's logs
- Filter Workloads by your Review App slug,
e.g.
review-qa-raise-e-12chm0
. - Find and open the
migrations
Deployment, e.g.review-qa-raise-e-12chm0-migrations.1
. - Click on the Pod in the "Managed pods" section, e.g.
review-qa-raise-e-12chm0-migrations.1-nqwtx
. - Click on the
Container logs
link.
Troubleshoot a pending dns-gitlab-review-app-external-dns
Deployment
Finding the problem
In the past, it happened
that the dns-gitlab-review-app-external-dns
Deployment was in a pending state,
effectively preventing all the Review Apps from getting a DNS record assigned,
making them unreachable via domain name.
This in turn prevented other components of the Review App to properly start
(e.g. gitlab-runner
).
After some digging, we found that new mounts were failing, when being performed
with transient scopes (e.g. pods) of systemd-mount
:
MountVolume.SetUp failed for volume "dns-gitlab-review-app-external-dns-token-sj5jm" : mount failed: exit status 1
Mounting command: systemd-run
Mounting arguments: --description=Kubernetes transient mount for /var/lib/kubelet/pods/06add1c3-87b4-11e9-80a9-42010a800107/volumes/kubernetes.io~secret/dns-gitlab-review-app-external-dns-token-sj5jm --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/06add1c3-87b4-11e9-80a9-42010a800107/volumes/kubernetes.io~secret/dns-gitlab-review-app-external-dns-token-sj5jm
Output: Failed to start transient scope unit: Connection timed out
This probably happened because the GitLab chart creates 67 resources, leading to a lot of mount points being created on the underlying GCP node.
The underlying issue seems to be a systemd
bug
that was fixed in systemd
v237
. Unfortunately, our GCP nodes are currently
using v232
.
For the record, the debugging steps to find out this issue were:
- Switch kubectl context to review-apps-ce (we recommend using kubectx)
kubectl get pods | grep dns
kubectl describe pod <pod name>
& confirm exact error message- Web search for exact error message, following rabbit hole to a relevant kubernetes bug report
- Access the node over SSH via the GCP console (Computer Engine > VM
instances then click the "SSH" button for the node where the
dns-gitlab-review-app-external-dns
pod runs) - In the node:
systemctl --version
=> systemd 232 - Gather some more information:
mount | grep kube | wc -l
=> e.g. 290systemctl list-units --all | grep -i var-lib-kube | wc -l
=> e.g. 142
- Check how many pods are in a bad state:
- Get all pods running a given node:
kubectl get pods --field-selector=spec.nodeName=NODE_NAME
- Get all the
Running
pods on a given node:kubectl get pods --field-selector=spec.nodeName=NODE_NAME | grep Running
- Get all the pods in a bad state on a given node:
kubectl get pods --field-selector=spec.nodeName=NODE_NAME | grep -v 'Running' | grep -v 'Completed'
- Get all pods running a given node:
Solving the problem
To resolve the problem, we needed to (forcibly) drain some nodes:
- Try a normal drain on the node where the
dns-gitlab-review-app-external-dns
pod runs so that Kubernetes automatically move it to another node:kubectl drain NODE_NAME
- If that doesn't work, you can also perform a forcible "drain" the node by removing all pods:
kubectl delete pods --field-selector=spec.nodeName=NODE_NAME
- In the node:
- Perform
systemctl daemon-reload
to remove the dead/inactive units - If that doesn't solve the problem, perform a hard reboot:
sudo systemctl reboot
- Perform
- Uncordon any cordoned nodes:
kubectl uncordon NODE_NAME
In parallel, since most Review Apps were in a broken state, we deleted them to
clean up the list of non-Running
pods.
Following is a command to delete Review Apps based on their last deployment date
(current date was June 6th at the time) with
helm ls -d | grep "Jun 4" | cut -f1 | xargs helm delete --purge
Mitigation steps taken to avoid this problem in the future
We've created a new node pool with smaller machines so that it's less likely that a machine will hit the "too many mount points" problem in the future.
Frequently Asked Questions
Isn't it too much to trigger CNG image builds on every test run? This creates thousands of unused Docker images.
We have to start somewhere and improve later. Also, we're using the CNG-mirror project to store these Docker images so that we can just wipe out the registry at some point, and use a new fresh, empty one.
How big are the Kubernetes clusters (review-apps-ce
and review-apps-ee
)?
The clusters are currently set up with a single pool of preemptible nodes, with a minimum of 1 node and a maximum of 500 nodes.
What are the machine running on the cluster?
We're currently using
n1-standard-1
(1 vCPU, 3.75 GB memory) machines.
How do we secure this from abuse? Apps are open to the world so we need to find a way to limit it to only us.
This isn't enabled for forks.