ctnr is a CLI built on top of runc
to manage and build OCI images as well as containers on Linux.
ctnr aims to ease system container creation and execution as unprivileged user.
Also ctnr is a tool to experiment with runc features.
THIS PROJECT IS NOT MAINTAINED ANYMORE IN FAVOUR OF podman.
- OCI bundle and container preparation as well as execution as unprivileged user using runc
- OCI image build as unprivileged user
- Simple concurrently accessible image and bundle store
- Image and bundle file system creation (based on umoci)
- Various image formats and transports supported by containers/image
- Container networking using CNI (optional, requires root, as OCI runtime hook)
- Dockerfile support
- Docker Compose 3 support (subset) using docker/cli (WIP)
- Easy to learn: docker-like CLI
- Easy installation: single statically linked binary (plus optional binaries: CNI plugins, proot) and convention over configuration
Concerning accessibility, usability and security a rootless container engine has several advantages:
- Containers can be run by unprivileged users.
Required in restrictive environments and useful for graphical applications. - Container images can be built in almost every Linux environment.
More flexibility in unprivileged builds - nesting containers is also possible (see experiments and limitations). - A higher degree and more flexible level of security.
Less likely for an attacker to gain root access when run as unprivileged user.
User/group-based container access control. Separation of responsibilities.
Container execution as unprivileged user is limited:
Container networking is limited.
With plain ctnr/runc only the host network can be used.
The standard CNI plugins require root privileges.
One workaround is to map ports on the host network using PRoot* accepting bad performance.
A better solution is to use slirp4netns which emulates the TCP/IP stack in a user namespace efficiently.
It can be used with ctnr via the slirp-cni-plugin.
Once container initialization is also moved into a user namespace with slirp the standard CNI plugins can be used again.
For instance the bridge can be used to achieve communication between containers (see user-mode networking).
Inside the container a process' or file's user cannot be changed.
This is caused by the fact that all operations in the container are still run by the host user (who is just mapped to user 0 inside the container).
Unfortunately this stops many package managers as well as official docker images from working:
While apk
or dnf
already work with plain runc apt-get
does not since it requires to change a user permanently.
To overcome this limitation ctnr supports the user.rootlesscontainers
xattr and integrates with PRoot*.
For more details see Aleksa Sarai's summary of the state of the art of rootless containers.
* PRoot is a binary that hooks its child processes' kernel-space system calls using ptrace
to simulate them in the user-space. This is more reliable but slower than hooking libc calls using LD_PRELOAD
as fakechroot does it.
Download the binary:
wget -O ctnr https://github.com/mgoltzsche/ctnr/releases/download/v0.7.0-alpha/ctnr.linux-amd64 &&
chmod +x ctnr &&
sudo mv ctnr /usr/local/bin/
If you need PRoot or CNI plugins
you can build them by calling make proot cni-plugins-static
within this repository's directory.
Build the binary dist/bin/ctnr
as well as dist/bin/cni-plugins
on a Linux machine with git, make and docker:
git clone https://github.com/mgoltzsche/ctnr.git
cd ctnr
make
Install in /usr/local
:
sudo make install
Optionally the project can now be opened with LiteIDE running in a ctnr container
(Please note that it takes some time to build the LiteIDE container image):
make ide
The following examples assume your policy accepts docker images or you have copied image-policy-example.json to /etc/containers/policy.json
on your host.
$ ctnr run docker://alpine:3.8 echo hello world
hello world
Build a Firefox ESR container image local/firefox:alpine
(cached operation):
$ ctnr image build \
--from=docker://alpine:3.8 \
--author='John Doe' \
--run='apk add --update --no-cache firefox-esr libcanberra-gtk3 adwaita-icon-theme ttf-ubuntu-font-family' \
--cmd=firefox \
--tag=local/firefox:alpine
Create and run a bundle named firefox
from the previously built image:
$ ctnr run -b firefox --update \
--env DISPLAY=$DISPLAY \
--mount src=/tmp/.X11-unix,dst=/tmp/.X11-unix \
--mount src=/etc/machine-id,dst=/etc/machine-id,opt=ro \
local/firefox:alpine
(Unfortunately tabs in firefox tend to crash)
The -b <BUNDLE>
and --update
options make this operation idempotent:
The bundle's file system is reused and only recreated when the underlying image has changed.
Use these options to restart containers very quickly. Without them ctnr copies the
image file system on bundle creation which can take some time and disk space depending on the image's size.
Also these options enable a container update on restart when the base image is frequently updated before the child image is rebuilt using the following command:
$ ctnr image import docker://alpine:3.8
This example shows how to build a debian-based image with the help of PRoot.
Dockerfile Dockerfile-cowsay
:
FROM debian:9
RUN apt-get update && apt-get install -y cowsay
ENTRYPOINT ["/usr/games/cowsay"]
Build the image (Please note that this works only with --proot
enabled. With plain ctnr/runc apt-get
fails to change uid/gid.):
$ ctnr image build --proot --dockerfile Dockerfile-cowsay --tag example/cowsay
Run a container using the previously built image (Please note that --proot
is not required anymore):
$ ctnr run example/cowsay hello from container
______________________
< hello from container >
----------------------
\ ^__^
\ (oo)\_______
(__)\ )\/\
||----w |
|| ||
ctnr supports port mapping using the -p, --publish
option.
Unprivileged users can use the --proot
option in addition.
When a container is run as root in a contained network (--network default
, default as root)
the portmap CNI plugin
is used to map ports from a specified IP or the host network to the container.
Map the container network's port 80 to port 8080 on the host:
$ sudo ctnr run -p 8080:80 docker://alpine:3.8 nc -l -p 80 -e echo hello from container
Connectivity test on the host on another shell:
$ nc 127.0.0.1 8080
hello from container
Unprivileged users can enable the --proot
option to map ports
within the host network namespace on a syscall level.
Map bind
/connect
syscalls with port 80 to port 8080:
$ ctnr run --proot -p 8080:80 docker://alpine:3.8 nc -l -p 80 -e echo hello from container
You can now also run another container using the same port as long as you don't map it on the same host port (proot maps it to a random free port and back within the container):
$ ctnr run --proot docker://alpine:3.8 /bin/sh -c 'nc -l -p 80 -e echo hello & sleep 1; timeout -t 1 nc 127.0.0.1 80'
hello
Connectivity test on the host on another shell:
$ nc 127.0.0.1 8080
hello from container
An OCI image provides a base configuration and file system to create an OCI bundle from. The file system consists of a list of layers represented by tar files each containing the diff to its predecessor.
ctnr manages images in its local store directory in the OCI image layout format.
Images are imported into the local store using the containers/image library.
A new bundle is created by extracting the image's file system into a directory and deriving the bundle's default configuration from the image's configuration plus user-defined options.
An OCI bundle describes a container by
a configuration and a file system.
Basically it is a directory containing a config.json
file with the configuration and a sub directory with the root file system.
ctnr manages bundles in its local store directory. Alternatively a custom directory can also be used as bundle.
OCI bundles generated by ctnr can also be run with plain runc.
An OCI container is a host-specific bundle instance.
On Linux it is a set of namespaces in which a configured process can be run.
ctnr provides two wrapper implementations of the OCI runtime reference implementation
runc/libcontainer
to either use an external runc binary or use libcontainer (no runtime dependencies!) controlled by a compiler flag.
- cri-o
- containerd
- docker
- lxc
- rkt
- rkt-compose
- runc
- runrootless
- singularity
- skopeo, umoci, orca-build
- udocker
- system.Context aware processes, unpacking/packing images
- improved multi-user support (store per user group, file permissions, lock location)
- CLI integration tests
- advanced rootless networking (using a network daemon run by root)
- separate OCI CNI network hook binary
- health check
- improved Docker Compose support
- service discovery integration (hook / DNS; consul, etcd)
- detached mode
- systemd integration (cgroup, startup notification)
- advanced logging
- support additional read-only image stores