Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add best practices for tooling #551

Merged
merged 8 commits into from
Aug 7, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions building/config.json
Original file line number Diff line number Diff line change
Expand Up @@ -717,6 +717,13 @@
"title": "Tooling Docker Setup",
"blurb": ""
},
{
"uuid": "83dc65b9-cc74-4a2a-9c70-472f6416c19d",
"slug": "tooling/best-practices",
"path": "building/tooling/best-practices.md",
"title": "Best Practices",
"blurb": ""
},
{
"uuid": "b080e814-d3b9-4027-a25e-907f5505cf8d",
"slug": "tooling/analyzers",
Expand Down
1 change: 1 addition & 0 deletions building/tooling/analyzers/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,3 +21,4 @@ You can use the following documents to learn more about building an analyzer:
- [Writing Analyzer comments](/docs/building/tooling/analyzers/comments)
- [Tagging solutions](/docs/building/tooling/analyzers/tags)
- [Guidance for building an Analyzer](/docs/building/tooling/analyzers/guidance)
- [Best practices](/docs/building/tooling/best-practices)
4 changes: 2 additions & 2 deletions building/tooling/analyzers/creating-from-scratch.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,9 @@ Firstly, thank you for your interest in creating an Analyzer!

These are the steps to get going:

1. Check [our repository list for an existing `...-analyzer`](https://github.com/exercism?q=-analyzer) to ensure that one doesn't already exist.
1. Check [our repository list for an existing `...-analyzer`](https://github.com/search?q=org%3Aexercism+analyzer&type=repositories) to ensure that one doesn't already exist.
2. Scan the [contents of this directory](/docs/building/tooling/analyzers) to ensure you are comfortable with the idea of creating an Analyzer.
3. Open an issue at [exercism/exercism][exercism-repo] introducing yourself and telling us which language you'd like to create a Analyzer for.
3. Start a new topic on [the Exercism forum][building-exercism] telling us which language you'd like to create an Analyzer for.
4. Once an Analyzer repo has been created, use [the Analyzer interface document](/docs/building/tooling/analyzers/interface) to help guide your implementation.

We have an incredibly friendly and supportive community who will be happy to help you as you work through this! If you get stuck, please start a new topic on [the Exercism forum][building-exercism] or create new issues at [exercism/exercism][exercism-repo] as needed 🙂
Expand Down
291 changes: 291 additions & 0 deletions building/tooling/best-practices.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,291 @@
# Best Practices

## Follow official best practices

The official [Dockerfile best practices](https://docs.docker.com/develop/develop-images/dockerfile_best-practices/) have lots of great content on how to improve your Dockerfiles.

## Performance

You should primarily optimize for performance (especially for test runners).
This will ensure your tooling runs as fast as possible and does not time-out.

### Experiment with different Base images

Try experimenting with different base images (e.g. Alpine instead of Ubuntu), to see if one (significantly) outperforms the other.
If performance is relatively equal, go for the image that is smallest.

### Try Internal Network

Check if using the `internal` network instead of `none` improves performance.
See the [network docs](/docs/building/tooling/docker#network) for more information.

### Prefer build-time commands over run-time commands

Tooling runs as one-off, short-lived Docker container:

1. A Docker container is created
2. The Docker container is run with the correct arguments
3. The Docker container is destroyed

Therefore, code that runs in step 2 runs for _every single tooling run_.
For this reason, reducing the amount of code that runs in step 2 is a great way to improve performance
One way of doing this is to move code from _run-time_ to _build-time_.
Whilst run-time code runs on every single tooling run, build-time code only runs once (when the Docker image is built).

Build-time code runs once as part of a GitHub Actions workflow.
Therefore, its fine if the code that runs at build-time is (relatively) slow.

#### Example: pre-compile libraries

When running tests in the Haskell test runner, it requires some base libraries to be compiled.
As each test run happens in a fresh container, this means that this compilation was done _in every single test run_!
To circumvent this, the [Haskell test runner's Dockerfile](https://github.com/exercism/haskell-test-runner/blob/5264c460054649fc672c3d5932c2f3cb082e2405/Dockerfile) has the following two commands:

```dockerfile
COPY pre-compiled/ .
RUN stack build --resolver lts-20.18 --no-terminal --test --no-run-tests
```

First, the `pre-compiled` directory is copied into the image.
This directory is setup as a sort of fake exercise and depends on the same base libraries that the actual exercise depend on.
Then we run the tests on that directory, which is similar to how tests are run for an actual exercise.
Running the tests will result in the base being compiled, but the difference is that this happens at _build time_.
The resulting Docker image will thus have its base libraries already compiled, which means that no longer has to happen at _run time_, resulting in (much) faster execution times.

#### Example: pre-compile binaries

Some languages allow code to be compiled ahead-of-time or just-in-time.
This is a build time vs. run time tradeoff, and again, we favor build time execution for performance reasons.

The [C# test runner's Dockerfile](https://github.com/exercism/csharp-test-runner/blob/b54122ef76cbf86eff0691daa33c8e50bc83979f/Dockerfile) uses this approach, where the test runner is compiled to a binary ahead-of-time (at build time) instead of just-in-time compiling the code (at run time).
This means that there is less work to do at run-time, which should help increase performance.

## Size

You should try to reduce the image's size, which means that it'll:

- Be faster to deploy
- Reduce costs for us
- Improve startup time of each container

### Try different distributions

Different distribution images will have different sizes.
For example, the `alpine:3.20.2` image is **ten times** smaller than the `ubuntu:24.10` image:

```
REPOSITORY TAG SIZE
alpine 3.20.2 8.83MB
ubuntu 24.10 101MB
```

In general, Alpine-based images are amongst the smallest images, so many tooling images are based on Alpine.

### Try slimmed-down images

Some images have special "slim" variants, in which some features will have been removed resulting in smaller image sizes.
For example, the `node:20.16.0-slim` image is **five times** smaller than the `node:20.16.0` image:

```
REPOSITORY TAG SIZE
node 20.16.0 1.09GB
node 20.16.0-slim 219MB
```

The reason "slim" variants are smaller is that they'll have less features.
Your image might not need the additional features, and if not, consider using the "slim" variant.

### Removing unneeded bits

An obvious, but great, way to reduce the size of your image is to remove anything you don't need.
These can include things like:

- Source files that are no longer needed after building a binary from them
- Files targeting different architectures from the Docker image
- Documentation

#### Remove package manager files

Most Docker images need to install additional packages, which is usually done via a package manager.
These packages must be installed at _build time_ (as no internet connection is available at _run time_).
Therefore, any package manager caching/bookkeeping files should be removed after installing the additional packages.

##### apk

Distributions that uses the `apk` package manager (such as Alpine) should use the `--no-cache` flag when using `apk add` to install packages:

```dockerfile
RUN apk add --no-cache curl
```

##### apt-get/apt

Distributions that use the `apt-get`/`apk` package manager (such as Ubuntu) should run the `apt-get autoremove -y` and `rm -rf /var/lib/apt/lists/*` commands _after_ installing the packages and in the same `RUN` command:

```dockerfile
RUN apt-get update && \
apt-get install curl -y && \
apt-get autoremove -y && \
rm -rf /var/lib/apt/lists/*
```

### Use multi-stage builds

Docker has a feature called [multi-stage builds](https://docs.docker.com/build/building/multi-stage/).
These allow you to partition your Dockerfile into separate _stages_, with only the last stage ending up in the produced Docker image (the rest is only there to support building the last stage).
You can think of each stage as its own mini Dockerfile; stages can use different base images.

Multi-stage builds are particularly useful when your Dockerfile requires packages to be installed that are _only_ needed at build time.
In this situation, the general structure of your Dockerfile looks like this:

1. Define a new stage (we'll call this the "build" stage).
This stage will _only_ be used at build time.
2. Install the required additional packages (into the "build" stage).
3. Run the commands that require the additional packages (within the "build" stage).
4. Define a new stage (we'll call this the "runtime" stage).
This stage will make up the resulting Docker image and executed at run time.
5. Copy the result(s) from the commands run in step 3 (in the "build" stage) into this stage (the "runtime" stage).

With this setup, the additional packages are _only_ installed in the "build" stage and _not_ in the "runtime" stage, which means that they won't end up in the Docker image that is produced.

#### Example: downloading files

The Fortran test runner requires `curl` to download some files.
However, its run time image does _not_ need `curl`, which makes this a perfect use case for a multi-stage build.

First, its [Dockerfile](https://github.com/exercism/fortran-test-runner/blob/783e228d8449143d2040e68b95128bb791833a27/Dockerfile) defines a stage (named "build") in which the `curl` package is installed.
It then uses curl to download files into that stage.

```dockerfile
FROM alpine:3.15 AS build

RUN apk add --no-cache curl

WORKDIR /opt/test-runner
COPY bust_cache .

WORKDIR /opt/test-runner/testlib
RUN curl -R -O https://raw.githubusercontent.com/exercism/fortran/main/testlib/CMakeLists.txt
RUN curl -R -O https://raw.githubusercontent.com/exercism/fortran/main/testlib/TesterMain.f90

WORKDIR /opt/test-runner
RUN curl -R -O https://raw.githubusercontent.com/exercism/fortran/main/config/CMakeLists.txt
```

The second part of the Dockerfile defines a new stage and copies the downloaded files from the "build" stage into its own stage using the `COPY` command:

```dockerfile
FROM alpine:3.15

RUN apk add --no-cache coreutils jq gfortran libc-dev cmake make

WORKDIR /opt/test-runner
COPY --from=build /opt/test-runner/ .

COPY . .
ENTRYPOINT ["/opt/test-runner/bin/run.sh"]
```

##### Example: installing libraries

The Ruby test runner needs the `git`, `openssh`, `build-base`, `gcc` and `wget` packages to be installed before its required libraries (gems) can be installed.
Its [Dockerfile](https://github.com/exercism/ruby-test-runner/blob/e57ed45b553d6c6411faeea55efa3a4754d1cdbf/Dockerfile) starts with a stage (given the name `build`) that install those packages (via `apk add`) and then installs the libaries (via `bundle install`):

```dockerfile
FROM ruby:3.2.2-alpine3.18 AS build

RUN apk update && apk upgrade && \
apk add --no-cache git openssh build-base gcc wget git

COPY Gemfile Gemfile.lock .

RUN gem install bundler:2.4.18 && \
bundle config set without 'development test' && \
bundle install
```

It then defines the stage that will form the resulting Docker image.
This stage does _not_ install the dependencies the previous stage installed, instead it uses the `COPY` command to copy the installed libraries from the build stage into its own stage:

```dockerfile
FROM ruby:3.2.2-alpine3.18

RUN apk add --no-cache bash

WORKDIR /opt/test-runner

COPY --from=build /usr/local/bundle /usr/local/bundle

COPY . .

ENTRYPOINT [ "sh", "/opt/test-runner/bin/run.sh" ]
```

```exercism/note
The [C# test runner's Dockerfile](https://github.com/exercism/csharp-test-runner/blob/b54122ef76cbf86eff0691daa33c8e50bc83979f/Dockerfile) does something similar, only in this case the build stage can use an existing Docker image that has pre-installed the additional packages required to install libraries.
```

## Safety

Safety is a main reason why we're using Docker containers to run our tooling.

### Prefer official images

There are many Docker images on [Docker Hub](https://hub.docker.com/), but try to use [official ones](https://hub.docker.com/search?q=&image_filter=official).
These images are curated and have (far) less chance of being unsafe.

### Pin versions

To ensure that builds are stable (i.e. they don't suddenly break), you should always pin your base images to specific tags.
That means instead of:

```dockerfile
FROM alpine:latest
```

you should use:

```dockerfile
FROM alpine:3.20.2
```

With the latter, builds will always use the same version.

### Run as a non-privileged user

By default, many images will run with a user that has root privileges.
You should consider running as a non-privileged user.

```dockerfile
FROM alpine

RUN groupadd -r myuser && useradd -r -g myuser myuser

# <RUN COMMANDS THAT REQUIRES ROOT USER, LIKE INSTALLING PACKAGES ETC.>

USER myuser
```

### Update package repositories to latest version

It is (almost) always a good idea to install the latest versions

```dockerfile
RUN apt-get update && \
apt-get install curl
```

### Support read-only filesystem

We encourage Docker files to be written using a read-only filesystem.
The only directories you should assume to be writeable are:

- The solution dir (passed in as the second argument)
- The output dir (passed in as the third argument)
- The `/tmp` dir

```exercism/caution
Our production environment currently does _not_ enforce a read-only filesystem, but we might in the future.
For this reason, the base template for a new test runner/analyzer/representer starts out with a read-only filesystem.
If you can't get things working on a read-only file, feel free to (for now) assume a writeable file system.
```
23 changes: 11 additions & 12 deletions building/tooling/docker.md
Original file line number Diff line number Diff line change
@@ -1,34 +1,32 @@
# Tooling Docker Setup
# Docker Setup

Our various track tooling are deployed as Docker images.

Each piece of tooling requires a Dockerfile, which specifies how the machine is built.
Each piece of tooling requires a [Dockerfile](https://docs.docker.com/reference/dockerfile/), which specifies how the machine is built.
It should live at the root directory of your repository and should be called `Dockerfile`.

The Dockerfile should create the minimal image needed for the tooling to function correctly and speedily.

The Dockerfile should produce an image with as a small a size as possible while maximizing (and prioritizing) performance.
Applying the official [Dockerfile best practices](https://docs.docker.com/develop/develop-images/dockerfile_best-practices/) can help to create a minimal image.
Our [Best Practices page](/docs/building/tooling/best-practices) has lots of tips to help you achieve this goal.

## Execution

### Timeouts

Each tooling run has a ten-second window in which to execute.
At the end of that period it will be timed out with a 408 error code.
The test runner gets 100% CPU with 3GB of memory for a 20 second window per solution.
After 20 seconds, the process is halted and reports a time-out with a 408 error code.

### Stdout/stderr

A tooling run may produce up to a maximum of one-megabyte of stdout and stderr.
If it produces more it will be killed with a 413 error code.
If it produces more, it will be killed with a 413 error code.

The contents of `stdout` and `stderr` from each run will be stored in files that can be viewed later.

You may write an `results.out` file to the output directory, which contains debugging information you want to later view.
You may write a `results.out` file to the output directory, which contains debugging information you want to view later.

### Results

The results file may be no larger than 500 kilobytes (including any stack traces etc).
The results file may not be larger than 500 kilobytes (including any stack traces etc).
If the file is larger than this, the tooling run will be killed with a 460 error code.

## Configuration
Expand All @@ -37,7 +35,8 @@ Each solution gets 100% machine resources for a twenty second window.

After 20 seconds, the process is halted and reports as a time-out.

Configuration can be set in the [`tools.json` file](https://github.com/exercism/tooling-invoker/blob/main/tools.json) in the Tooling Invoker repository.
Some tools require (slight) deviations from the default configuration.
If so, these are configured in the [`tools.json` file](https://github.com/exercism/tooling-invoker/blob/main/tools.json) in the Tooling Invoker repository.

### Network

Expand Down Expand Up @@ -67,4 +66,4 @@ docker container run -v /path/to/job:/mnt/exercism-iteration --network none -m 1

## Editing a Dockerfile

All changes to Dockerfiles require a PR review from the @exercism/ops team, in order to avoid the introduction of security exploits.
All changes to Dockerfiles require a PR review from the `@exercism/maintainers-admin` team, in order to avoid the introduction of security exploits.
1 change: 1 addition & 0 deletions building/tooling/representers/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -35,3 +35,4 @@ You can use the following documents to learn more about building a representer:
- [The Representer interface](/docs/building/tooling/representers/interface)
- [How to normalize representations for the highest efficiency](/docs/building/tooling/representers/normalization)
- [How to build a Docker image with Docker for local testing and deployment](/docs/building/tooling/representers/docker)
- [Best practices](/docs/building/tooling/best-practices)
4 changes: 2 additions & 2 deletions building/tooling/representers/creating-from-scratch.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,9 @@ Firstly, thank you for your interest in creating a Representer!

These are the steps to get going:

1. Check [our repository list for an existing `...-representer`](https://github.com/exercism?q=-representer) to ensure that one doesn't already exist.
1. Check [our repository list for an existing `...-representer`](https://github.com/search?q=org%3Aexercism+representer&type=repositories) to ensure that one doesn't already exist.
2. Scan the [contents of this directory](/docs/building/tooling/representers) to ensure you are comfortable with the idea of creating an Representer.
3. Open an issue at [exercism/exercism][exercism-repo] introducing yourself and telling us which language you'd like to create a Representer for.
3. Start a new topic on [the Exercism forum][building-exercism] telling us which language you'd like to create a Representer for.
4. Once a Representer repo has been created, use [the Representer interface document](/docs/building/tooling/representers/interface) to help guide your implementation.

We have an incredibly friendly and supportive community who will be happy to help you as you work through this! If you get stuck, please start a new topic on [the Exercism forum][building-exercism] or create new issues at [exercism/exercism][exercism-repo] as needed 🙂
Expand Down
Loading
Loading