Go to file
Priya Wadhwa c63a03d123
rename to kaniko
2018-04-04 14:52:13 -07:00
deploy Merged master 2018-04-04 13:35:20 -07:00
executor Merged master, fixed merge conflict 2018-04-04 14:38:39 -07:00
files Add auth for runing in Kubernetes clusteer 2018-03-12 16:01:45 -07:00
hack hack: check dep against correct upstream 2018-03-06 16:21:10 -08:00
integration_tests rename to kaniko 2018-04-04 14:52:13 -07:00
kaniko Add README and update name to kaniko 2018-04-02 14:00:15 -07:00
pkg rename to kaniko 2018-04-04 14:52:13 -07:00
testutil Copy command and unit tests 2018-03-14 17:06:46 -07:00
vendor Merged master, fixed conflict 2018-04-04 14:06:38 -07:00
.gcloudignore Integration tests 2018-02-22 12:06:23 -08:00
.gitignore Initial skeleton for executor (#22) 2018-02-20 19:05:54 -08:00
.travis.yml Initial skeleton for executor (#22) 2018-02-20 19:05:54 -08:00
CONTRIBUTING.md Initial commit 2018-01-25 14:22:07 -08:00
Gopkg.lock Merged master, fixed conflict 2018-04-04 14:06:38 -07:00
Gopkg.toml Unpack filesystem and whitelist from /proc/self/mountinfo 2018-02-21 11:02:30 -08:00
LICENSE Initial commit 2018-01-25 14:22:07 -08:00
Makefile Merged master 2018-04-04 13:35:20 -07:00
README.md rename to kaniko 2018-04-04 14:52:13 -07:00
integration-test.sh Integration tests 2018-02-22 12:06:23 -08:00
run_in_docker.sh Merged master 2018-04-04 13:35:20 -07:00
test.sh Unpack filesystem and whitelist from /proc/self/mountinfo 2018-02-21 11:02:30 -08:00

README.md

kaniko

kaniko is a tool to build unpriviliged container images from a Dockerfile. It doesn't depend on a Docker daemon, which enables building container images in environments that can't easily or securely run a Docker daemon, such as a standard Kubernetes cluster.

The majority of Dockerfile commands can be executed with kaniko, but we're still working on supporting the following commands:

  • VOLUME
  • SHELL
  • HEALTHCHECK
  • STOPSIGNAL
  • ONBUILD
  • ARG

We're currently in the process of building kaniko, so as of now it isn't production ready. Please let us know if you have any feature requests or find any bugs!

How does kaniko work?

The kaniko executor image is responsible for building an image from a Dockerfile and pushing it to a registry. Within the executor image, we extract the filesystem of the base image (the FROM image in the Dockerfile). We then execute the commands in the Dockerfile, snapshotting the filesystem in userspace after each one. After each command, we append a layer of changed files to the base image (if there are any) and update image metadata.

kaniko Build Contexts

kaniko supports local directories and GCS buckets as build contexts. To specify a local directory, pass in the --context flag as an argument to the executor image. To specify a GCS bucket, pass in the --bucket flag. The GCS bucket should contain a compressed tar of the build context called context.tar.gz, which kaniko will unpack and use as the build context.

To easily create context.tar.gz, we can use skaffold. Running the following command within the build context will create context.tar.gz, which will contain the Dockerfile and any files it depends on.

skaffold docker context

We can copy over the compressed tar to a GCS bucket with gsutil:

gsutil cp context.tar.gz gs://<bucket name>

Running kaniko locally

Requirements:

  • Docker
  • gcloud

We can run the kaniko executor image locally in a Docker daemon to build and push an image from a Dockerfile.

First, we want to load the executor image into the Docker daemon by running

make images

To run kaniko in Docker, run the following command:

./run_in_docker.sh <path to Dockerfile> <path to build context> <destination of final image>

Running kaniko in a Kubernetes cluster

Requirements:

  • Standard Kubernetes cluster
  • Kubernetes Secret

To run kaniko in a Kubernetes cluster, you will need a standard running Kubernetes cluster and a Kubernetes secret, which contains the auth required to push the final image.

To create the secret, first you will need to create a service account in the Pantheon project you want to push the final image to, with Storage Admin permissions. You can download a JSON key for this service account, and rename it kaniko-secret.json. To create the secret, run:

kubectl create secret generic kaniko-secret --from-file=<path to kaniko-secret.json>

The Kubernetes job.yaml should look similar to this, with the args parameters filled in:

apiVersion: batch/v1
kind: Job
metadata:
  name: kaniko
spec:
  template:
    spec:
      containers:
      - name: kaniko
        image: gcr.io/kaniko-project/executor:latest
        args: ["--dockerfile=<path to Dockerfile>", "--bucket=<GCS bucket>", "--destination=<gcr.io/$PROJECT/$IMAGE:$TAG>"]
        volumeMounts:
          - name: kaniko-secret
            mountPath: /secret
        env:
          - name: GOOGLE_APPLICATION_CREDENTIALS
            value: /secret/kaniko-secret.json
      restartPolicy: Never
      volumes:
        - name: kaniko-secret
          secret:
            secretName: kaniko-secret

This example pulls the build context from a GCS bucket. To use a local directory build context, you could consider using configMaps to mount in small build contexts.

Comparison with Other Tools

Similar tools include:

All of these tools build container images with different approaches. Both kaniko and img build unprivileged images, but they interpret “unprivileged” differently. img builds as a non root user from within the container, while kaniko is run in an unprivileged environment with root access inside the container.

orca-build depends on runC to build images from Dockerfiles; since kaniko doesn't use runC it doesn't require the use of kernel namespacing techniques.

buildah requires the same root privilges as a Docker daemon does to run, while kaniko runs without any special privileges or permissions.

FTL aims to achieve the fastest possible creation of Docker images for a subset of images. It can be thought of as a special-case "fast path" that can be used in conjunction with the support for general Dockerfiles kaniko provides.