Allow to select packaged plugins in gerrit-master chart

The gerrit-master chart installed some selected plugins packaged in the
gerrit.war on startup, but did not allow to select which core plugins
should be installed.

This change adds an option to the values.yaml chart, which allows to
hand over a list of packaged plugins to install for Gerrit. The default
value is the fixed list of plugins, which was previously installed:

    - commit-message-length-validator
    - download-commands
    - replication
    - reviewnotes

Change-Id: I88900e62824f27e51da72387dc7630b360cb5da8
8 files changed
tree: 7a10322fbd1351df32eba89f7289ebfc920dd92a
  1. .github/
  2. container-images/
  3. Documentation/
  4. helm-charts/
  5. supplements/
  6. tests/
  7. .gitignore
  8. .pylintrc
  9. build
  10. LICENSE
  11. Pipfile
  12. Pipfile.lock
  13. publish
  14. README.md
  15. setup.cfg
README.md

Gerrit Deployment on Kubernetes

Container images, configurations and Helm charts for installing Gerrit on Kubernetes.

Docker images

Images to run a Gerrit master and slave setup based on the latest stable-2.12 Gerrit build.

Building images

To build all images, the build-script in the root directory of the project can be used:

./build

If a specific image should be build, the image name can be specified as an argument. Multiple images can be specified at once:

./build gerrit-slave git-gc

The build-script usually uses the latest-tag to tag the images. By using the --tag TAG-option, a custom tag can be defined:

./build --tag test

The build script will in addition tag the image with the output of git describe --dirty.

The single component images inherit a base image. The Dockerfile for the base image can be found in the ./base-directory. It will be automatically built by the ./build-script. If the component images are built manually, the base image has to be built first with the target base:latest, since it is not available in a registry and thus has to exist locally.

Publishing images

The publish script in the root directory of the project can be used to push the built images to the configured registry. To do so, log in first, before executing the script.

docker login <registry>

To configure the registry and image version, the respective values can be configured via env variables REGISTRY and TAG. In addition, these values can also be passed as command line options named --registry and --tag in which case they override the values from env variables:

./publish <component-name>

The <component-name> is one of: apache-git-http-backend, git-gc, gerrit-slave.

Adding the --update-latest-flag will also update the images tagged latest in the repository:

./publish --update-latest <component-name>

Running images in Docker

The container images are meant to be used by the helm-charts provided in this project. The images are thus not designed to be used in a standalone setup. To run Gerrit on Docker use the docker-gerrit project.

Important notes

Currently, java is installed under /usr/lib/jvm/java-8-openjdk-amd64/jre. Therefore, make sure that container.javaHome is set to that path in the gerrit.config:

  javaHome = /usr/lib/jvm/java-8-openjdk-amd64/jre

Helm Charts

These Helm charts can be used to install a Gerrit cluster consisting of a Gerrit master and a Gerrit slave on a Kubernetes cluster.

To evaluate and test the helm-charts, they can be installed on a local machine running Minikube. Follow this guide to get a detailed description how to set up the Minikube cluster and install the charts.

File System Storage

Currently this deployment uses NFS, some options:

  • Create an EFS volume on AWS
  • Install a NFS server on Kubernetes cluster which doesn't have read-write-many Persistent Volumes available using NFS-provisioner

Gerrit Master

Gerrit Slave

Running tests

The tests are implemented using Python and pytest. To ensure a well-defined test-environment, pipenv is meant to be used to install packages and provide a virtual environment in which to run the tests. To install pipenv, use brew:

brew install pipenv

More detailed information can be found in the pipenv GitHub repo.

To create the virtual environment with all required packages, run:

pipenv install

To run all tests, execute:

pipenv run pytest

Some tests will need to create files in a temporary directory. Some of these files will be mounted into docker containers by tests. For this to work make either sure that the system temporary directory is accessible by the Docker daemon or set the base temporary directory to a directory accessible by Docker by executing:

pipenv run pytest --basetemp=/tmp/k8sgerrit

By default the tests will build all images from scratch. This will greatly increase the time needed for testing. To use already existing container images, a tag can be provided as follows:

pipenv run pytest --tag=v0.1

The tests will then use the existing images with the provided tag. If an image does not exist, it will still be built by the tests.

By default the build of the container images will not use the build cache created by docker. To enable the cache, execute:

pipenv run pytest --build-cache

Slow tests may be marked with the decorator @pytest.mark.slow. These tests may then be skipped as follows:

pipenv run pytest --skip-slow

There are also other marks, allowing to select tests (refer to this section).

To run specific tests, execute one of the following:

# Run all tests in a directory (including subdirectories)
pipenv run pytest tests/container-images/base

# Run all tests in a file
pipenv run pytest tests/container-images/base/test_container_build_base.py

# Run a specific test
pipenv run \
  pytest tests/container-images/base/test_container_build_base.py::test_build_base

# Run tests with a specific marker
pipenv run pytest -m "docker"

For a more detailed description of how to use pytest, refer to the official documentation.

Test marks

docker

Marks tests which start up docker containers. These tests will interact with the containers by either using docker exec or sending HTTP-requests. Make sure that your system supports this kind of interaction.

incremental

Marks test classes in which the contained test functions have to run incrementally.

integration

Marks integration tests. These tests test interactions between containers, between outside clients and containers and between the components installed by a helm chart.

kubernetes

Marks tests that require a Kubernetes cluster. These tests are used to test the functionality of the helm charts in this project and the interaction of the components installed by them. The cluster should not be used for other purposes to minimize unforeseen interactions.

Currently these tests also require access to an EFS volume.

slow

Marks tests that need an above average time to run.

structure

Marks structure tests. These tests are meant to test, whether certain components exist in a container. These tests ensure that components expected by the users of the container, e.g. the helm charts, are present in the containers.