Images to run a Gerrit master and slave setup based on the 3.0.3 Gerrit release.
To build all images, the
build-script in the root directory of the project can be used:
If a specific image should be build, the image name can be specified as an argument. Multiple images can be specified at once:
./build gerrit-slave git-gc
The build-script usually uses the
latest-tag to tag the images. By using the
--tag TAG-option, a custom tag can be defined:
./build --tag test
The build script will in addition tag the image with the output of
git describe --dirty.
The single component images inherit a base image. The
Dockerfile for the base image can be found in the
./base-directory. It will be automatically built by the
./build-script. If the component images are built manually, the base image has to be built first with the target
base:latest, since it is not available in a registry and thus has to exist locally.
The publish script in the root directory of the project can be used to push the built images to the configured registry. To do so, log in first, before executing the script.
docker login <registry>
To configure the registry and image version, the respective values can be configured via env variables
TAG. In addition, these values can also be passed as command line options named
--tag in which case they override the values from env variables:
<component-name> is one of:
--update-latest-flag will also update the images tagged
latest in the repository:
./publish --update-latest <component-name>
The container images are meant to be used by the helm-charts provided in this project. The images are thus not designed to be used in a standalone setup. To run Gerrit on Docker use the docker-gerrit project.
Currently, java is installed under
/usr/lib/jvm/java-8-openjdk-amd64/jre. Therefore, make sure that
container.javaHome is set to that path in the
javaHome = /usr/lib/jvm/java-8-openjdk-amd64/jre
These Helm charts can be used to install a Gerrit cluster consisting of a Gerrit master and a Gerrit slave on a Kubernetes cluster.
To evaluate and test the helm-charts, they can be installed on a local machine running Minikube. Follow this guide to get a detailed description how to set up the Minikube cluster and install the charts.
Currently this deployment uses NFS, some options:
The tests are implemented using Python and
pytest. To ensure a well-defined test-environment,
pipenv is meant to be used to install packages and provide a virtual environment in which to run the tests. To install pipenv, use
brew install pipenv
More detailed information can be found in the pipenv GitHub repo.
To create the virtual environment with all required packages, run:
To run all tests, execute:
pipenv run pytest -m "not smoke"
-m "not smoke"-option excludes the smoke tests, which will fail, since no Gerrit-instance will be running, when they are executed.
Some tests will need to create files in a temporary directory. Some of these files will be mounted into docker containers by tests. For this to work make either sure that the system temporary directory is accessible by the Docker daemon or set the base temporary directory to a directory accessible by Docker by executing:
pipenv run pytest --basetemp=/tmp/k8sgerrit -m "not smoke"
By default the tests will build all images from scratch. This will greatly increase the time needed for testing. To use already existing container images, a tag can be provided as follows:
pipenv run pytest --tag=v0.1 -m "not smoke"
The tests will then use the existing images with the provided tag. If an image does not exist, it will still be built by the tests.
By default the build of the container images will not use the build cache created by docker. To enable the cache, execute:
pipenv run pytest --build-cache -m "not smoke"
Slow tests may be marked with the decorator
@pytest.mark.slow. These tests may then be skipped as follows:
pipenv run pytest --skip-slow -m "not smoke"
There are also other marks, allowing to select tests (refer to this section).
To run specific tests, execute one of the following:
# Run all tests in a directory (including subdirectories) pipenv run pytest tests/container-images/base # Run all tests in a file pipenv run pytest tests/container-images/base/test_container_build_base.py # Run a specific test pipenv run \ pytest tests/container-images/base/test_container_build_base.py::test_build_base # Run tests with a specific marker pipenv run pytest -m "docker"
For a more detailed description of how to use
pytest, refer to the official documentation.
Marks tests which start up docker containers. These tests will interact with the containers by either using
docker exec or sending HTTP-requests. Make sure that your system supports this kind of interaction.
Marks test classes in which the contained test functions have to run incrementally.
Marks integration tests. These tests test interactions between containers, between outside clients and containers and between the components installed by a helm chart.
Marks tests that require a Kubernetes cluster. These tests are used to test the functionality of the helm charts in this project and the interaction of the components installed by them. The cluster should not be used for other purposes to minimize unforeseen interactions.
Currently these tests also require access to an EFS volume.
Marks tests that need an above average time to run.
Marks structure tests. These tests are meant to test, whether certain components exist in a container. These tests ensure that components expected by the users of the container, e.g. the helm charts, are present in the containers.
To run smoke tests, use the following command:
pipenv run pytest \ -m "smoke" \ --basetemp="<tmp-dir for tests>" \ --ingress-url="<Gerrit URL>" \ --gerrit-user="<Gerrit user>" \ --gerrit-pwd
The smoke tests require a Gerrit user that is allowed to create and delete projects. The username has to be given by
--gerit-user. Setting the
--gerrit-pwd-flag will cause a password prompt to enter the password of the Gerrit-user.