This document describes how to configure runtime options for
containerd-shim-runsc-v1
. You can find the installation instructions and
minimal requirements in Containerd Quick Start.
The shim can be provided with a configuration file containing options to the shim itself as well as a set of flags to runsc. Here is a quick example:
cat <<EOF | sudo tee /etc/containerd/runsc.toml
option = "value"
[runsc_config]
flag = "value"
EOF
The set of options that can be configured can be found in
options.go.
Values under [runsc_config]
can be used to set arbitrary flags to runsc.
flag = "value"
is converted to --flag="value"
when runsc is invoked. Run
runsc flags
so see which flags are available
Next, containerd needs to be configured to send the configuration file to the shim.
Starting in 1.3, containerd supports a configurable ConfigPath
in the runtime
configuration. Here is an example:
cat <<EOF | sudo tee /etc/containerd/config.toml
version = 2
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
runtime_type = "io.containerd.runc.v2"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc]
runtime_type = "io.containerd.runsc.v1"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc.options]
TypeUrl = "io.containerd.runsc.v1.options"
ConfigPath = "/etc/containerd/runsc.toml"
EOF
When you are done, restart containerd to pick up the changes.
sudo systemctl restart containerd
When shim_debug
is enabled in /etc/containerd/config.toml
, containerd will
forward shim logs to its own log. You can additionally set level = "debug"
to
enable debug logs. To see the logs run sudo journalctl -u containerd
. Here is
a containerd configuration file that enables both options:
cat <<EOF | sudo tee /etc/containerd/config.toml
version = 2
[debug]
level = "debug"
[plugins."io.containerd.runtime.v1.linux"]
shim_debug = true
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
runtime_type = "io.containerd.runc.v2"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc]
runtime_type = "io.containerd.runsc.v1"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc.options]
TypeUrl = "io.containerd.runsc.v1.options"
ConfigPath = "/etc/containerd/runsc.toml"
EOF
It can be hard to separate containerd messages from the shim’s though. To create
a log file dedicated to the shim, you can set the log_path
and log_level
values in the shim configuration file:
log_path
is the directory where the shim logs will be created. %ID%
is
the path is replaced with the container ID.log_level
sets the logs level. It is normally set to “debug” as there is
not much interesting happening with other log levels.gVisor debug logging can be enabled by setting the debug
and debug-log
flag.
The shim will replace “%ID%” with the container ID, and “%COMMAND%” with the
runsc command (run, boot, etc.) in the path of the debug-log
flag.
Find out more about debugging in the debugging guide.
cat <<EOF | sudo tee /etc/containerd/runsc.toml
log_path = "/var/log/runsc/%ID%/shim.log"
log_level = "debug"
[runsc_config]
debug = "true"
debug-log = "/var/log/runsc/%ID%/gvisor.%COMMAND%.log"
EOF
If you want to use
nvidia-container-runtime
with runsc through containerd, you might need to configure nvidia
runtime in
containerd via sudo nvidia-ctk runtime configure --runtime=containerd
command.
This will update /etc/containerd/config.toml
with a new runtime named
nvidia
. However, this runtime’s configuration is not compatible with runsc:
runtime_type
is set to runc. You will need to manually update this
field to specify runsc so that containerd tries to invoke
containerd-shim-runsc-v1
when using nvidia
runtime.options
attempts to specify BinaryName =
"/usr/bin/nvidia-container-runtime"
. However, runsc shim takes
configuration via ConfigPath
as shown above. So the options
needs to be
updated to specify ConfigPath
and in the config.toml file needs to specify
the BinaryName
.The /etc/containerd/config.toml
file should look like:
...
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.nvidia]
runtime_type = "io.containerd.runsc.v1"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.nvidia.options]
TypeUrl = "io.containerd.runsc.v1.options"
ConfigPath = "/etc/containerd/runsc.toml"
...
And /etc/containerd/runsc.toml
should look something like:
log_path = "/var/log/runsc/%ID%/shim.log"
log_level = "debug"
binary_name = "/usr/bin/nvidia-container-runtime"
[runsc_config]
debug = "true"
debug-log = "/var/log/runsc/%ID%/gvisor.%COMMAND%.log"
nvproxy = "true"
See this section for information about
configuring nvidia-container-runtime
to use runsc
as its low-level runtime.
When multiple containers share the same volume in a gVisor sandbox, inotify may fail to detect cross-container file changes unless specific annotations and configuration are set. By default, gVisor mounts container volumes independently without a single view of the entire pod. As a result, changes to shared volumes from one container appear as external updates to other containers. While these updates may be detected upon access, inotify does not properly pick them up unless extra metadata (i.e., “mount hints”) is provided.
GKE often handles these hints automatically in its control plane, but on other Kubernetes distributions (e.g., EKS), you must configure them manually. Below are the steps and annotations required.
In your /etc/containerd/config.toml
, configure containerd to propagate custom
gVisor annotations to runsc:
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.gvisor]
runtime_type = "io.containerd.runsc.v1"
pod_annotations = [ "dev.gvisor.*" ]
Mount hints must be provided in the form of annotations that tell gVisor how to handle each volume:
dev.gvisor.spec.mount.<NAME>.share
: container
, pod
, or shared
.
container
: The volume is only used by one container.pod
: The volume is used by multiple containers within the pod.shared
: The volume is shared with outside the pod, and requires
frequent checks for external changes.dev.gvisor.spec.mount.<NAME>.type
: tmpfs
or bind
.
tmpfs
: The volume uses tmpfs
inside the sandbox. For emptyDir
volumes, setting this to tmpfs
enables better performance.bind
: Indicates a bind mount from the host.dev.gvisor.spec.mount.<NAME>.options
: Comma-separated volume-mount options
(e.g., rw,rprivate
).Below is an example Pod spec for a shared emptyDir
volume named
shared-folder
, with annotations that enable cross-container inotify:
apiVersion: v1
kind: Pod
metadata:
name: shared-folder-test
annotations:
dev.gvisor.spec.mount.shared-folder.share: "pod"
dev.gvisor.spec.mount.shared-folder.type: "tmpfs"
dev.gvisor.spec.mount.shared-folder.options: "rw,rprivate"
spec:
runtimeClassName: gvisor
containers:
- name: container1
image: node:14
command: ["node", "watcher.js"]
volumeMounts:
- name: shared-folder
mountPath: /shared
- name: container2
image: busybox
command: ["/bin/sh", "-c", "while true; do echo 'hello' > /shared/test.txt; sleep 2; done"]
volumeMounts:
- name: shared-folder
mountPath: /shared
volumes:
- name: shared-folder
emptyDir: {}
dev.gvisor.spec.mount.<NAME>.share
: pod
indicates that multiple
containers within the same sandbox share this volume. This sets up fast
access and allows inotify events to propagate across containers.dev.gvisor.spec.mount.<NAME>.type
: tmpfs
is often recommended when
using emptyDir
to enable higher performance inside the sandbox. Although
this is referred to as a tmpfs
, it may still be backed by disk if
configured that way.dev.gvisor.spec.mount.<NAME>.options
: Typically rw,rprivate
for a
shared read/write volume.Once these annotations are in place, and the Pod is scheduled with the gvisor
runtime, inotify should correctly detect file changes across containers sharing
the same volume.
If you have trouble, verify:
pod_annotations
) is in place.shared-folder
) matches the annotation keys.With these configurations, shared volumes in multi-container pods using gVisor on EKS (or other environments) should behave similarly to GKE by enabling cross-container inotify-based file change detection.