Containerd Advanced Configuration

Edit this page Create issue

This document describes how to configure runtime options for containerd-shim-runsc-v1. You can find the installation instructions and minimal requirements in Containerd Quick Start.

Shim Configuration

The shim can be provided with a configuration file containing options to the shim itself as well as a set of flags to runsc. Here is a quick example:

cat <<EOF | sudo tee /etc/containerd/runsc.toml
option = "value"
[runsc_config]
  flag = "value"
EOF

The set of options that can be configured can be found in options.go. Values under [runsc_config] can be used to set arbitrary flags to runsc. flag = "value" is converted to --flag="value" when runsc is invoked. Run runsc flags so see which flags are available

Next, containerd needs to be configured to send the configuration file to the shim.

Containerd 1.3+

Starting in 1.3, containerd supports a configurable ConfigPath in the runtime configuration. Here is an example:

cat <<EOF | sudo tee /etc/containerd/config.toml
version = 2
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
  runtime_type = "io.containerd.runc.v2"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc]
  runtime_type = "io.containerd.runsc.v1"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc.options]
  TypeUrl = "io.containerd.runsc.v1.options"
  ConfigPath = "/etc/containerd/runsc.toml"
EOF

When you are done, restart containerd to pick up the changes.

sudo systemctl restart containerd

Debug

When shim_debug is enabled in /etc/containerd/config.toml, containerd will forward shim logs to its own log. You can additionally set level = "debug" to enable debug logs. To see the logs run sudo journalctl -u containerd. Here is a containerd configuration file that enables both options:

cat <<EOF | sudo tee /etc/containerd/config.toml
version = 2
[debug]
  level = "debug"
[plugins."io.containerd.runtime.v1.linux"]
  shim_debug = true
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
  runtime_type = "io.containerd.runc.v2"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc]
  runtime_type = "io.containerd.runsc.v1"
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runsc.options]
  TypeUrl = "io.containerd.runsc.v1.options"
  ConfigPath = "/etc/containerd/runsc.toml"
EOF

It can be hard to separate containerd messages from the shim’s though. To create a log file dedicated to the shim, you can set the log_path and log_level values in the shim configuration file:

  • log_path is the directory where the shim logs will be created. %ID% is the path is replaced with the container ID.
  • log_level sets the logs level. It is normally set to “debug” as there is not much interesting happening with other log levels.

Example: Enable shim and gVisor debug logging

gVisor debug logging can be enabled by setting the debug and debug-log flag. The shim will replace “%ID%” with the container ID, and “%COMMAND%” with the runsc command (run, boot, etc.) in the path of the debug-log flag.

Find out more about debugging in the debugging guide.

cat <<EOF | sudo tee /etc/containerd/runsc.toml
log_path = "/var/log/runsc/%ID%/shim.log"
log_level = "debug"
[runsc_config]
  debug = "true"
  debug-log = "/var/log/runsc/%ID%/gvisor.%COMMAND%.log"
EOF

NVIDIA Container Runtime

If you want to use nvidia-container-runtime with runsc through containerd, you might need to configure nvidia runtime in containerd via sudo nvidia-ctk runtime configure --runtime=containerd command. This will update /etc/containerd/config.toml with a new runtime named nvidia. However, this runtime’s configuration is not compatible with runsc:

  • Its runtime_type is set to runc. You will need to manually update this field to specify runsc so that containerd tries to invoke containerd-shim-runsc-v1 when using nvidia runtime.
  • Its options attempts to specify BinaryName = "/usr/bin/nvidia-container-runtime". However, runsc shim takes configuration via ConfigPath as shown above. So the options needs to be updated to specify ConfigPath and in the config.toml file needs to specify the BinaryName.

The /etc/containerd/config.toml file should look like:

...
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.nvidia]
  runtime_type = "io.containerd.runsc.v1"

[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.nvidia.options]
  TypeUrl = "io.containerd.runsc.v1.options"
  ConfigPath = "/etc/containerd/runsc.toml"
...

And /etc/containerd/runsc.toml should look something like:

log_path = "/var/log/runsc/%ID%/shim.log"
log_level = "debug"
binary_name = "/usr/bin/nvidia-container-runtime"
[runsc_config]
  debug = "true"
  debug-log = "/var/log/runsc/%ID%/gvisor.%COMMAND%.log"
  nvproxy = "true"

See this section for information about configuring nvidia-container-runtime to use runsc as its low-level runtime.

Enabling inotify for Shared Volumes

When multiple containers share the same volume in a gVisor sandbox, inotify may fail to detect cross-container file changes unless specific annotations and configuration are set. By default, gVisor mounts container volumes independently without a single view of the entire pod. As a result, changes to shared volumes from one container appear as external updates to other containers. While these updates may be detected upon access, inotify does not properly pick them up unless extra metadata (i.e., “mount hints”) is provided.

GKE often handles these hints automatically in its control plane, but on other Kubernetes distributions (e.g., EKS), you must configure them manually. Below are the steps and annotations required.

1. Allow Pod Annotations to Pass to gVisor

In your /etc/containerd/config.toml, configure containerd to propagate custom gVisor annotations to runsc:

[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.gvisor]
  runtime_type = "io.containerd.runsc.v1"
  pod_annotations = [ "dev.gvisor.*" ]

2. Add gVisor Volume-Mount Annotations

Mount hints must be provided in the form of annotations that tell gVisor how to handle each volume:

  • dev.gvisor.spec.mount.<NAME>.share: container, pod, or shared.
    • container: The volume is only used by one container.
    • pod: The volume is used by multiple containers within the pod.
    • shared: The volume is shared with outside the pod, and requires frequent checks for external changes.
  • dev.gvisor.spec.mount.<NAME>.type: tmpfs or bind.
    • tmpfs: The volume uses tmpfs inside the sandbox. For emptyDir volumes, setting this to tmpfs enables better performance.
    • bind: Indicates a bind mount from the host.
  • dev.gvisor.spec.mount.<NAME>.options: Comma-separated volume-mount options (e.g., rw,rprivate).

Below is an example Pod spec for a shared emptyDir volume named shared-folder, with annotations that enable cross-container inotify:

apiVersion: v1
kind: Pod
metadata:
  name: shared-folder-test
  annotations:
    dev.gvisor.spec.mount.shared-folder.share: "pod"
    dev.gvisor.spec.mount.shared-folder.type: "tmpfs"
    dev.gvisor.spec.mount.shared-folder.options: "rw,rprivate"
spec:
  runtimeClassName: gvisor
  containers:
    - name: container1
      image: node:14
      command: ["node", "watcher.js"]
      volumeMounts:
        - name: shared-folder
          mountPath: /shared
    - name: container2
      image: busybox
      command: ["/bin/sh", "-c", "while true; do echo 'hello' > /shared/test.txt; sleep 2; done"]
      volumeMounts:
        - name: shared-folder
          mountPath: /shared
  volumes:
    - name: shared-folder
      emptyDir: {}

Explanation of Annotations

  • dev.gvisor.spec.mount.<NAME>.share: pod indicates that multiple containers within the same sandbox share this volume. This sets up fast access and allows inotify events to propagate across containers.
  • dev.gvisor.spec.mount.<NAME>.type: tmpfs is often recommended when using emptyDir to enable higher performance inside the sandbox. Although this is referred to as a tmpfs, it may still be backed by disk if configured that way.
  • dev.gvisor.spec.mount.<NAME>.options: Typically rw,rprivate for a shared read/write volume.

Once these annotations are in place, and the Pod is scheduled with the gvisor runtime, inotify should correctly detect file changes across containers sharing the same volume.

If you have trouble, verify:

  • That the containerd config snippet (pod_annotations) is in place.
  • The volume name (e.g., shared-folder) matches the annotation keys.
  • Your gVisor debug logs for warnings or errors.

With these configurations, shared volumes in multi-container pods using gVisor on EKS (or other environments) should behave similarly to GKE by enabling cross-container inotify-based file change detection.