Using image mode for RHEL to build, deploy, and manage operating systems
Using RHEL bootc images on Red Hat Enterprise Linux 9
Abstract
Providing feedback on Red Hat documentation
We appreciate your feedback on our documentation. Let us know how we can improve it.
Submitting feedback through Jira (account required)
- Log in to the Jira website.
- Click Create in the top navigation bar
- Enter a descriptive title in the Summary field.
- Enter your suggestion for improvement in the Description field. Include links to the relevant parts of the documentation.
- Click Create at the bottom of the dialogue.
Chapter 1. Introducing image mode for RHEL
Use image mode for RHEL to build, test, and deploy operating systems by using the same tools and techniques as application containers. Image mode for RHEL is available by using the registry.redhat.io/rhel9/rhel-bootc
bootc image. The RHEL bootc images differ from the existing application Universal Base Images (UBI) in that they contain additional components necessary to boot that were traditionally excluded, such as, kernel, initrd, boot loader, firmware, among others.
Red Hat provides the rhel9/rhel-9-bootc
container image as a Technology Preview. Technology Preview features provide early access to upcoming product innovations, enabling customers to test functionality and provide feedback during the development process. However, these features are not fully supported. Documentation for a Technology Preview feature might be incomplete or include only basic installation and configuration information. See Technology Preview Features Support Scope on the Red Hat Customer Portal for information about the support scope for Technology Preview features.
The rhel-bootc
and user-created containers based on rhel-bootc
container image are subject to the Red Hat Enterprise Linux end user license agreement (EULA). You are not allowed to publicly redistribute these images.
Figure 1.1. Building, deploying, and managing operating system by using image mode for RHEL

Red Hat provides bootc image for the following computer architectures:
- AMD and Intel 64-bit architectures (x86-64-v2)
- The 64-bit ARM architecture (ARMv8.0-A)
The benefits of image mode for RHEL occur across the lifecycle of a system. The following list contains some of the most important advantages:
- Container images are easier to understand and use than other image formats and are fast to build
- Containerfiles, also known as Dockerfiles, provide a straightforward approach to defining the content and build instructions for an image. Container images are often significantly faster to build and iterate on compared to other image creation tools.
- Consolidate process, infrastructure, and release artifacts
- As you distribute applications as containers, you can use the same infrastructure and processes to manage the underlying operating system.
- Immutable updates
-
Just as containerized applications are updated in an immutable way, with image mode for RHEL, the operating system is also. You can boot into updates and roll back when needed in the same way that you use
rpm-ostree
systems.
The use of rpm-ostree
to make changes, or install content, is not supported.
- Portability across hybrid cloud environments
- You can use bootc images across physical, virtualized, cloud, and edge environments.
Although containers provide the foundation to build, transport, and run images, it is important to understand that after you deploy these bootc images, either by using an installation mechanism, or you convert them to a disk image, the system does not run as a container.
- Bootc supports the following container image formats and disk image formats:
Image type | Target environment |
---|---|
| Physical, virtualized, cloud, and edge environments. |
| Amazon Machine Image. |
| QEMU. |
| VMDK for vSphere. |
| An unattended Anaconda installer that installs to the first disk found. |
| Unformatted raw disk. Also supported in QEMU and Libvirt |
| VHD for Virtual PC, among others. |
| Google Compute Engine (GCE) environment. |
Containers help streamline the lifecycle of a RHEL system by offering the following possibilities:
- Building container images
-
You can configure your operating system at a build time by modifying the Containerfile. Image mode for RHEL is available by using the
registry.redhat.io/rhel9/rhel-bootc
container image. You can use Podman, OpenShift Container Platform, or other standard container build tools to manage your containers and container images. You can automate the build process by using CI/CD pipelines. - Versioning, mirroring, and testing container images
- You can version, mirror, introspect, and sign your derived bootc image by using any container tools such as Podman or OpenShift Container Platform.
- Deploying container images to the target environment
You have several options on how to deploy your image:
- Anaconda: is the installation program used by RHEL. You can deploy all image types to the target environment by using Anaconda and Kickstart to automate the installation process.
-
bootc-image-builder
: is a containerized tool that converts the container image to different types of disk images, and optionally uploads them to an image registry or object storage. -
bootc
: is a tool responsible for fetching container images from a container registry and installing them to a system, updating the operating system, or switching from an existing ostree-based system. The RHEL bootc image contains thebootc
utility by default and works with all image types. However, remember that therpm-ostree
is not supported and must not be used to make changes.
- Updating your operating system
-
The system supports in-place transactional updates with rollback after deployment. Automatic updates are on by default. A systemd service unit and systemd timer unit files check the container registry for updates and apply them to the system. As the updates are transactional, a reboot is required. For environments that require more sophisticated or scheduled rollouts, disable auto updates and use the
bootc
utility to update your operating system.
RHEL has two deployment modes. Both provide the same stability, reliability, and performance during deployment. See their differences:
-
Package mode: You can build package-based images and OSTree images by using RHEL image builder, and you can manage the package mode images by using
composer-cli
or web console. The operating system uses RPM packages and is updated by using thednf
package manager. The root filesystem is mutable. However, the operating system cannot be managed as a containerized application. See Composing a customized RHEL system image product documentation. -
Image mode: a container-native approach to build, deploy, and manage RHEL. The same RPM packages are delivered as a base image and updates are deployed as a container image. The root filesystem is immutable by default, except for
/etc
and/var
, with most content coming from the container image.
You can choose to use either the Image mode or the Package mode deployment to build, test, and share your operating system. Image mode additionally enables you to manage your operating system in the same way as any other containerized application.
1.1. Prerequisites
- You have a subscribed RHEL 9 system. For more information, see Getting Started with RHEL System Registration documentation.
- You have a container registry. You can create your registry locally or create a free account on the Quay.io service. To create the Quay.io account, see Red Hat Quay.io page.
- You have a Red Hat account with either production or developer subscriptions. No cost developer subscriptions are available on the Red Hat Enterprise Linux Overview page.
- You have authenticated to registry.redhat.io. For more information, see Red Hat Container Registry Authentication article.
1.2. Additional resources
- Introducing image mode for RHEL and bootc in Podman Desktop quick start guide
- Image mode for Red Hat Enterprise Linux quick start: AI inference quick start guide
- Getting Started with Podman AI Lab blog article
- Customizing Anaconda product documentation
- Automatically installing RHEL product documentation (Kickstart)
- Composing a customized RHEL system image product documentation
- Composing, installing, and managing RHEL for Edge images product documentation
Chapter 2. Building and testing RHEL bootc images
The following procedures use Podman to build and test your container image. You can also use other tools, for example, OpenShift Container Platform. For more examples of configuring RHEL systems by using containers, see the rhel-bootc-examples repository.
Figure 2.1. Building an image by using instructions from a Containerfile, testing the container, pushing an image to a registry, and sharing it with others

A general Containerfile
structure is the following:
FROM registry.redhat.io/rhel9/rhel-bootc:latest RUN dnf -y install [software] [dependencies] && dnf clean all ADD [application] ADD [configuration files] RUN [config scripts]
The available commands that are usable inside a Containerfile
and a Dockerfile
are equivalent.
However, the following commands in a Containerfile
are ignored when the rhel-9-bootc
image is installed to a system:
-
ENTRYPOINT
andCMD
(OCI:Entrypoint/Cmd
): you can setCMD /sbin/init
instead. -
ENV
(OCI:Env
): change thesystemd
configuration to configure the global system environment. -
EXPOSE
(OCI:exposedPorts
): it is independent of how the system firewall and network function at runtime. -
USER
(OCI:User
): configure individual services inside the RHEL bootc to run as unprivileged users instead.
The rhel-9-bootc
container image reuses the OCI image format.
-
The
rhel-9-bootc
container image ignores the container config section (Config
) when it is installed to a system. -
The
rhel-9-bootc
container image does not ignore the container config section (Config
) when you run this image by using container runtimes such aspodman
ordocker
.
Building custom rhel-bootc
base images is not supported in this release.
2.1. Building a container image
Use the podman build
command to build an image using instructions from a Containerfile
.
Prerequisites
-
The
container-tools
meta-package is installed.
Procedure
Create a
Containerfile
:$ cat Containerfile FROM registry.redhat.io/rhel9/rhel-bootc:latest RUN dnf -y install cloud-init && \ ln -s ../cloud-init.target /usr/lib/systemd/system/default.target.wants && \ dnf clean all
This
Containerfile
example adds thecloud-init
tool, so it automatically fetches SSH keys and can run scripts from the infrastructure and also gather configuration and secrets from the instance metadata. For example, you can use this container image for pre-generated AWS or KVM guest systems.Build the
<image>
image by usingContainerfile
in the current directory:$ podman build -t quay.io/<namespace>/<image>:<tag> .
Verification
List all images:
$ podman images REPOSITORY TAG IMAGE ID CREATED SIZE localhost/<image> latest b28cd00741b3 About a minute ago 2.1 GB
Additional resources
2.2. Building derived bootable images by using multi-stage builds
The deployment image should include only the application and its required runtime, without adding any build tools or unnecessary libraries. To achieve this, use a two-stage Containerfile
: one stage for building the artifacts and another for hosting the application.
With multi-stage builds, you use multiple FROM
instructions in your Containerfile
. Each FROM
instruction can use a different base, and each of them begins a new stage of the build. You can selectively copy artifacts from one stage to another, and exclude everything you do not need in the final image.
Multi-stage builds offer several advantages:
- Smaller image size
- By separating the build environment from the runtime environment, only the necessary files and dependencies are included in the final image, significantly reducing its size.
- Improved security
- Since build tools and unnecessary libraries are excluded from the final image, the attack surface is reduced, leading to a more secure container.
- Optimized performance
- A smaller image size means faster download, deployment, and startup times, improving the overall efficiency of the containerized application.
- Simplified maintenance
- With the build and runtime environments separated, the final image is cleaner and easier to maintain, containing only what is needed to run the application.
- Cleaner builds
- Multi-stage builds help avoid clutter from intermediate files, which could accumulate during the build process, ensuring that only essential artifacts make it into the final image.
- Resource efficiency
- The ability to build in one stage and discard unnecessary parts minimizes the use of storage and bandwidth during deployment.
- Better Layer Caching
- With clearly defined stages, Podman can efficiently cache the results of previous stages, by accelerating up future builds.
The following Containerfile
consists of two stages. The first stage is typically named builder
and it compiles a golang binary. The second stage copies the binary from the first stage. The default working directory for the go-toolset builder is opt/ap-root/src
.
FROM registry.access.redhat.com/ubi9/go-toolset:latest as builder RUN echo 'package main; import "fmt"; func main() { fmt.Println("hello world") }' > helloworld.go RUN go build helloworld.go FROM registry.redhat.io/rhel9/rhel-bootc:latest COPY --from=builder /opt/app-root/src/helloworld / CMD ["/helloworld"]
As a result, the final container image includes the helloworld
binary but no data from the previous stage.
You can also use multi-stage builds to perform the following scenarios:
- Stopping at a specific build stage
- When you build your image, you can stop at a specified build stage. For example:
$ podman build --target build -t hello .
For example, you can use this approach to debugging a specific build stage.
- Using an external image as a stage
-
You can use the
COPY --from
instruction to copy from a separate image either using the local image name, a tag available locally or on a container registry, or a tag ID. For example:
COPY --from=<image> <source_path> <destination_path>_
- Using a previous stage as a new stage
-
You can continue where a previous stage ended by using the
FROM
instruction. From example:
FROM ubi9 AS stage1 [...] FROM stage1 AS stage2 [...] FROM ubi9 AS final-stage [...]
Additional resources
2.3. Running a container image
Use the podman run
command to run and test your container.
Prerequisites
-
The
container-tools
meta-package is installed.
Procedure
Run the container named
mybootc
based on thequay.io/<namespace>/<image>:<tag>
container image:$ podman run -it --rm --name mybootc quay.io/<namespace>/<image>:<tag> /bin/bash
-
The
-i
option creates an interactive session. Without the-t
option, the shell stays open, but you cannot type anything to the shell. -
The
-t
option opens a terminal session. Without the-i
option, the shell opens and then exits. -
The
--rm
option removes thequay.io/<namespace>/<image>:<tag>
container image after the container exits.
-
The
Verification
List all running containers:
$ podman ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 7ccd6001166e quay.io/<namespace>/<image>:<tag> /sbin/init 6 seconds ago Up 5 seconds ago mybootc
Additional resources
2.4. Pushing a container image to the registry
Use the podman push
command to push an image to your own, or a third party, registry and share it with others. The following procedure uses the Red Hat Quay registry.
Prerequisites
-
The
container-tools
meta-package is installed. - An image is built and available on the local system.
- You have created the Red Hat Quay registry. For more information see Proof of Concept - Deploying Red Hat Quay.
Procedure
Push the
quay.io/<namespace>/<image>:<tag>
container image from your local storage to the registry:$ podman push quay.io/<namespace>/<image>:<tag>
Additional resources
-
podman-tag(1)
man page -
podman-push(1)
man page
Chapter 3. Building and managing logically bound images
Logically bound images give you support for container images that are lifecycle bound to the base bootc image. This helps combine different operational processes for applications and operating systems, and the container application images are referenced from the base image as image files or an equivalent. Consequently, you can manage multiple container images for system installations.
You can use containers for lifecycle-bound workloads, such as security agents and monitoring tools. You can also upgrade such workloads by using the bootc upgrade
command.
3.1. Logically bound images
Logically bound images enable an association of the container application images to a base bootc system image. The term logically bound is used to contrast with physically bound images. The logically bound images offer the following benefits:
- You can update the bootc system without re-downloading the application container images.
- You can update the application container images without modifying the bootc system image, which is especially useful for development work.
The following are examples for lifecycle bound workloads, whose activities are usually not updated outside of the host:
- Logging, for example, journald→remote log forwarder container
- Monitoring, for example, Prometheus node_exporter
- Configuration management agents
- Security agents
Another important property of the logically bound images is that they must be present and available on the host, possibly from a very early stage in the boot process.
Differently from the default usage of tools like Podman or Docker, images might be pulled dynamically after the boot starts, which requires a functioning network. For example, if the remote registry is temporarily unavailable, the host system might run longer without log forwarding or monitoring, which is not desirable. Logically bound images enable you to reference container images similarly to you can with ExecStart=
in a systemd unit.
When using logically bound images, you must manage multiple container images for the system to install the logically bound images. This is an advantage and also a disadvantage. For example, for a disconnected or offline installation, you must mirror all the containers, not just one. The application images are only referenced from the base image as .image
files or an equivalent.
- Logically bound images installation
-
When you run
bootc install
, logically bound images must be present in the default/var/lib/containers
container store. The images will be copied into the target system and present directly at boot, alongside the bootc base image. - Logically bound images lifecycle
-
Logically bound images are referenced by the bootable container and have guaranteed availability when the bootc based server starts. The image is always upgraded by using
bootc upgrade
and is available asread-only
to other processes, such as Podman. - Logically bound images management on upgrades, rollbacks, and garbage collection
- During upgrades, the logically bound images are managed exclusively by bootc.
- During rollbacks, the logically bound images corresponding to rollback deployments are retained.
- bootc performs garbage collection of unused bound images.
3.2. Using logically bound images
Each logically bound image is defined in a Podman Quadlet .image
or .container file
. To use logically bound images, follow the steps:
Prerequisites
-
The
container-tools
meta-package is installed.
Procedure
- Select the image that you want to logically bound.
Create a
Containerfile
:$ cat Containerfile FROM quay.io/<namespace>/<image>:latest COPY ./<app_1>.image /usr/share/containers/systemd/<app_1>.image COPY ./<app_2>.container /usr/share/containers/systemd/<app_2>.container RUN ln -s /usr/share/containers/systemd/<app_1>.image \ /usr/lib/bootc/bound-images.d/<app_1>.image && \ ln -s /usr/share/containers/systemd/<app_2>.container \ /usr/lib/bootc/bound-images.d/<app_2>.container
In the
.container
definition, use:GlobalArgs=--storage-opt=additionalimagestore=/usr/lib/bootc/storage
In the
Containerfile
example, the image is selected to be logically bound by creating a symlink in the/usr/lib/bootc/bound-images.d
directory pointing to either an.image
or a.container
file.Run the
bootc upgrade
command.$ bootc upgrade
The bootc upgrade performs the following overall steps:
- Fetches the new base image from the image repository. See linkhttps://docs.redhat.com/en/documentation/red_hat_enterprise_linux/9/html/using_image_mode_for_rhel_to_build_deploy_and_manage_operating_systems/managing-users-groups-ssh-key-and-secrets-in-image-mode-for-rhel#configuring-container-pull-secrets_managing-users-groups-ssh-key-and-secrets-in-image-mode-for-rhel[Configuring container pull secrets].
- Reads the new base image root file system to discover logically bound images.
-
Automatically pulls any discovered logically bound images defined in the new bootc image into the bootc-owned
/usr/lib/bootc/storage
image storage.
Make the bound images become available to container runtimes such as Podman. For that, you must explicitly configure bound images to point to the bootc storage as an "additional image store". For example:
podman --storage-opt=additionalimagestore=/usr/lib/bootc/storage run <image>
ImportantDo not attempt to globally enable the
/usr/lib/bootc/storage
image storage in/etc/containers/storage.conf
. Only use the bootc storage for logically bound images.
The bootc image store
is owned by bootc
. The logically bound images will be garbage collected when they are no longer referenced by a file in the /usr/lib/bootc/bound-images.d
directory.
Chapter 4. Creating bootc compatible base disk images with bootc-image-builder
The bootc-image-builder
, available as a Technology Preview, is a containerized tool to create disk images from bootc images. You can use the images that you build to deploy disk images in different environments, such as the edge, server, and clouds.
4.1. Introducing image mode for RHEL for bootc-image-builder
With the bootc-image-builder
tool, you can convert bootc images into disk images for a variety of different platforms and formats. Converting bootc images into disk images is equivalent to installing a bootc. After you deploy these disk images to the target environment, you can update them directly from the container registry.
The bootc-image-builder
can only pull and use images from public container repositories. Building base disk images which come from private registries by using bootc-image-builder
is not supported in this release. If your container image is stored in a private repository, bootc-image-builder
cannot pull the image because it is not able to authenticate to the registry. If you need to use an image from a private repository, you must authenticate to the registry first and then pull the container image before you use it with bootc-image-builder
. After pulling the image, you can run the bootc-image-builder command using the --local
option.
The bootc-image-builder tool
supports generating the following image types:
- Disk image formats, such as ISO, suitable for disconnected installations.
Virtual disk images formats, such as:
- QEMU copy-on-write (QCOW2)
- Amazon Machine Image (AMI)/ — Raw
- Virtual Machine Image (VMI)
Deploying from a container image is beneficial when you run VMs or servers because you can achieve the same installation result. That consistency extends across multiple different image types and platforms when you build them from the same container image. Consequently, you can minimize the effort in maintaining operating system images across platforms. You can also update systems that you deploy from these disk images by using the bootc
tool, instead of re-creating and uploading new disk images with bootc-image-builder
.
Generic base container images do not include any default passwords or SSH keys. Also, the disk images that you create by using the bootc-image-builder
tool do not contain the tools that are available in common disk images, such as cloud-init
. These disk images are transformed container images only.
Although you can deploy a rhel-9-bootc
image directly, you can also create your own customized images that are derived from this bootc image. The bootc-image-builder
tool takes the rhel-9-bootc
OCI container image as an input.
Additional resources
4.2. Installing bootc-image-builder
The bootc-image-builder
is intended to be used as a container and it is not available as an RPM package in RHEL. To access it, follow the procedure.
Prerequisites
-
The
container-tools
meta-package is installed. The meta-package contains all container tools, such as Podman, Buildah, and Skopeo. -
You are authenticated to
registry.redhat.io
. For details, see Red Hat Container Registry Authentication.
Procedure
Login to authenticate to
registry.redhat.io
:$ sudo podman login registry.redhat.io
Install the
bootc-image-builder
tool:$ sudo podman pull registry.redhat.io/rhel9/bootc-image-builder
Verification
List all images pulled to your local system:
$ sudo podman images REPOSITORY TAG IMAGE ID CREATED SIZE registry.redhat.io/rhel9/bootc-image-builder latest b361f3e845ea 24 hours ago 676 MB
Additional resources
4.3. Creating QCOW2 images by using bootc-image-builder
Build a RHEL bootc image into a QEMU Disk Images (QCOW2) image for the architecture that you are running the commands on.
The RHEL base image does not include a default user. Optionally, you can inject a user configuration with the --config
option to run the bootc-image-builder container. Alternatively, you can configure the base image with cloud-init
to inject users and SSH keys on first boot. See Users and groups configuration - Injecting users and SSH keys by using cloud-init.
Prerequisites
- You have Podman installed on your host machine.
-
You have
virt-install
installed on your host machine. -
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Optional: Create a
config.toml
to configure user access, for example:[[customizations.user]] name = "user" password = "pass" key = "ssh-rsa AAA ... user@email.com" groups = ["wheel"]
Run
bootc-image-builder
. Optionally, if you want to use user access configuration, pass theconfig.toml
as an argument.NoteIf you do not have the container storage mount and
--local
image options, your image must be public.The following is an example of creating a public QCOW2 image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v ./config.toml:/config.toml \ -v ./output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type qcow2 \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private QCOW2 image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type qcow2 \ quay.io/<namespace>/<image>:<tag>
You can find the
.qcow2
image in the output folder.
Next steps
- You can deploy your image. See Deploying a container image using KVM with a QCOW2 disk image.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
4.4. Creating VMDK images by using bootc-image-builder
Create a Virtual Machine Disk (VMDK) from a bootc image and use it within VMware’s virtualization platforms, such as vSphere, or use the Virtual Machine Disk (VMDK) in VirtualBox.
Prerequisites
- You have Podman installed on your host machine.
-
You have authenticated to the Red Hat Registry by using the
podman login registry.redhat.io
. -
You have pulled the
rhel9/bootc-image-builder
container image.
Procedure
Create a
Containerfile
with the following content:FROM registry.redhat.io/rhel9/rhel-bootc:9.4 RUN dnf -y install cloud-init open-vm-tools && \ ln -s ../cloud-init.target /usr/lib/systemd/system/default.target.wants && \ rm -rf /var/{cache,log} /var/lib/{dnf,rhsm} && \ systemctl enable vmtoolsd.service
Build the bootc image:
# podman build . -t localhost/rhel-bootc-vmdk
Create a VMDK file from the previously created bootc image:
NoteIf you do not have the container storage mount and --local image options, your image must be public.
The following is an example of creating a public VMDK image:
# podman run \ --rm \ -it \ --privileged \ -v /var/lib/containers/storage:/var/lib/containers/storage \ -v ./output:/output \ --security-opt label=type:unconfined_t \ --pull newer \ registry.redhat.io/rhel9/bootc-image-builder:9.4 --local \ --type vmdk \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private VMDK image:
# podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type vmdk \ quay.io/<namespace>/<image>:<tag>
The
--local
option uses the local container storage to source the originating image to produce the VMDK instead of a remote repository.
A VMDK disk file for the bootc image is stored in the output/vmdk
directory.
Next steps
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
4.5. Creating GCE images by using bootc-image-builder
Build a RHEL bootc image into a gce image for the architecture that you are running the commands on. The RHEL base image does not include a default user. Optionally, you can inject a user configuration with the --config
option to run the bootc-image-builder container. Alternatively, you can configure the base image with cloud-init
to inject users and SSH keys on first boot. See Users and groups configuration - Injecting users and SSH keys by using cloud-init.
Prerequisites
- You have Podman installed on your host machine.
-
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Optional: Create a
config.toml
to configure user access, for example:[[customizations.user]] name = "user" password = "pass" key = "ssh-rsa AAA ... user@email.com" groups = ["wheel"]
Run
bootc-image-builder
. Optionally, if you want to use user access configuration, pass theconfig.toml
as an argument.NoteIf you do not have the container storage mount and
--local
image options, your image must be public.The following is an example of creating a
gce
image:$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v ./config.toml:/config.toml \ -v ./output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type gce \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private
gce
image:$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type gce \ quay.io/<namespace>/<image>:<tag>
You can find the
gce
image in the output folder.
Next steps
- You can deploy your image. See Deploying a container image using KVM with a QCOW2 disk image.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
4.6. Creating AMI images by using bootc-image-builder and uploading it to AWS
Create an Amazon Machine Image (AMI) from a bootc image and use it to launch an Amazon Web Service EC2 (Amazon Elastic Compute Cloud) instance.
Prerequisites
- You have Podman installed on your host machine.
-
You have an existing
AWS S3
bucket within your AWS account. -
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images. -
You have the
vmimport
service role configured on your account to import an AMI into your AWS account.
Procedure
Create a disk image from the bootc image.
- Configure the user details in the Containerfile. Make sure that you assign it with sudo access.
- Build a customized operating system image with the configured user from the Containerfile. It creates a default user with passwordless sudo access.
Optional: Configure the machine image with
cloud-init
. See Users and groups configuration - Injecting users and SSH keys by using cloud-init. The following is an example:FROM registry.redhat.io/rhel9/rhel-bootc:9.4 RUN dnf -y install cloud-init && \ ln -s ../cloud-init.target /usr/lib/systemd/system/default.target.wants && \ rm -rf /var/{cache,log} /var/lib/{dnf,rhsm}
NoteYou can also use
cloud-init
to add users and additional configuration by using instance metadata.Build the bootc image. For example, to deploy the image to an
x86_64
AWS machine, use the following commands:$ podman build -t quay.io/<namespace>/<image>:<tag> . $ podman push quay.io/<namespace>/<image>:<tag> .
Use the
bootc-image-builder
tool to create an AMI from the bootc container image.NoteIf you do not have the container storage mount and
--local
image options, your image must be public.The following is an example of creating a public AMI image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ -v $HOME/.aws:/root/.aws:ro \ --env AWS_PROFILE=default \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type ami \ --aws-ami-name rhel-bootc-x86 \ --aws-bucket rhel-bootc-bucket \ --aws-region us-east-1 \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private AMI image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ -v $HOME/.aws:/root/.aws:ro \ --env AWS_PROFILE=default \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type ami \ --aws-ami-name rhel-bootc-x86 \ --aws-bucket rhel-bootc-bucket \ --local quay.io/<namespace>/<image>:<tag>
You can also inject all your AWS configuration parameters by using
--env AWS_*
.ImportantThe following flags must be specified all together. If you do not specify any flag, the AMI is exported to your output directory.
-
--aws-ami-name
- The name of the AMI image in AWS -
--aws-bucket
- The target S3 bucket name for intermediate storage when you are creating the AMI --aws-region
- The target region for AWS uploadsThe
bootc-image-builder
tool builds an AMI image and uploads it to your AWS s3 bucket by using your AWS credentials to push and register an AMI image after building it.
-
Next steps
- You can deploy your image. See Deploying a container image to AWS with an AMI disk image.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
Additional resources
4.7. Creating raw disk images by using bootc-image-builder
You can convert a bootc image to a Raw image with an MBR or GPT partition table by using bootc-image-builder
. The RHEL base image does not include a default user, so optionally, you can inject a user configuration with the --config
option to run the bootc-image-builder
container. Alternatively, you can configure the base image with cloud-init
to inject users and SSH keys on first boot. See Users and groups configuration - Injecting users and SSH keys by using cloud-init.
Prerequisites
- You have Podman installed on your host machine.
-
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images. - You have pulled your target container image in the container storage.
Procedure
Optional: Create a
config.toml
to configure user access, for example:[[customizations.user]] name = "user" password = "pass" key = "ssh-rsa AAA ... user@email.com" groups = ["wheel"]
Run
bootc-image-builder
. If you want to use user access configuration, pass theconfig.toml
as an argument:NoteIf you do not have the container storage mount and
--local
image options, your image must be public.The following is an example of creating a public RAW image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v /var/lib/containers/storage:/var/lib/containers/storage \ -v ./config.toml:/config.toml \ -v ./output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local \ --type raw \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private RAW image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type raw \ quay.io/<namespace>/<image>:<tag>
You can find the
.raw
image in the output folder.
Next steps
- You can deploy your image. See Deploying a container image by using KVM with a QCOW2 disk image.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
4.8. Creating ISO images by using bootc-image-builder
You can use bootc-image-builder
to create an ISO from which you can perform an offline deployment of a bootable container.
Prerequisites
- You have Podman installed on your host machine.
-
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Optional: Create a
config.toml
to configure user access, for example:[[customizations.user]] name = "user" password = "pass" key = "ssh-rsa AAA ... user@email.com" groups = ["wheel"]
Run
bootc-image-builder
. If you do not want to add any configuration, omit the-v $(pwd)/config.toml:/config.toml
argument.NoteIf you do not have the container storage mount and --local image options, your image must be public.
The following is an example of creating a public ISO image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v /var/lib/containers/storage:/var/lib/containers/storage \ -v $(pwd)/config.toml:/config.toml \ -v $(pwd)/output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type iso \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private ISO image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type iso \ quay.io/<namespace>/<image>:<tag>
You can find the
.iso
image in the output folder.
Next steps
You can use the ISO image on unattended installation methods, such as USB sticks or Install-on-boot. The installable boot ISO contains a configured Kickstart file. See Deploying a container image by using Anaconda and Kickstart.
WarningBooting the ISO on a machine with an existing operating system or data can be destructive, because the Kickstart is configured to automatically reformat the first disk on the system.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootable images.
4.9. Using bootc-image-builder to build ISO images with a Kickstart file
You can use a Kickstart file to configure various parts of the installation process, such as setting up users, customizing partitioning, and adding an SSH key. You can include the Kickstart file in an ISO build to configure any part of the installation process, except the deployment of the base image. For ISOs with bootc container base images, you can use a Kickstart file to configure anything except the ostreecontainer
command.
For example, you can use a Kickstart to perform either a partial installation, a full installation, or even omit the user creation. Use bootc-image-builder
to build an ISO image that contains the custom Kickstart to configure your installation process.
Prerequisites
- You have Podman installed on your host machine.
-
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Create your Kickstart file. The following Kickstart file is an example of a fully unattended Kickstart file configuration that contains user creation, and partition instructions.
[customizations.installer.kickstart] contents = """ lang en_GB.UTF-8 keyboard uk timezone CET user --name <user> --password <password> --plaintext --groups <groups> sshkey --username <user> ssh-<type> <public key> rootpw --lock zerombr clearpart --all --initlabel autopart --type=plain reboot --eject """
-
Save the Kickstart configuration in the
toml
format to inject the Kickstart content. For example,config.toml
. Run
bootc-image-builder
, and include the Kickstart file configuration that you want to add to the ISO build. Thebootc-image-builder
automatically adds theostreecontainer
command that installs the container image.NoteIf you do not have the container storage mount and
--local
image options, your image must be public.The following is an example of creating a public ISO image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v /var/lib/containers/storage:/var/lib/containers/storage \ -v $(pwd)/config.toml:/config.toml \ -v $(pwd)/output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type iso \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
The following is an example of creating a private ISO image:
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type iso \ quay.io/<namespace>/<image>:<tag>
You can find the
.iso
image in the output folder.
Next steps
You can use the ISO image on unattended installation methods, such as USB sticks or Install-on-boot. The installable boot ISO contains a configured Kickstart file. See Deploying a container image by using Anaconda and Kickstart.
WarningBooting the ISO on a machine with an existing operating system or data can be destructive, because the Kickstart is configured to automatically reformat the first disk on the system.
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootable images.
4.10. Verification and troubleshooting
- If you have any issues configuring the requirements for your AWS image, see the following documentation
- For more details on users, groups, SSH keys, and secrets, see
Chapter 5. Customizing disk images of RHEL image mode with advanced partitioning
There are 2 options to customize advanced partitions:
- Disk customizations
- Filesystem customizations
However, the two customizations are incompatible with each other. You cannot use both customizations in the same blueprint.
5.1. Understanding partitions
The following are the general principles about partitions:
- The full disk image size is always larger than the size of the sum of the partitions, due to requirements for headers and metadata. Consequently, all sizes are treated as minimum requirements, whether for specific filesystems, partitions, logical volumes, or the image itself.
- When the partition is automatically added, the partition that contains the root filesystem is always the last in the partition table layout. This is valid for a plain formatted partition, an LVM Volume Group, or a Btrfs partition. For disk customizations, the order that you defined is respected.
-
For the raw partitioning, that is, with no LVM, the partition containing the root filesystem is grown to fill any leftover space on the partition table. Logical Volumes are not grown to fill the space in the Volume Group because they are simple to grow on a live system. Some directories have hard-coded minimum sizes which cannot be overridden. These are 1 GiB for
/
and 2 GiB for/usr
. As a result, if/usr
is not on a separate partition, the root filesystem size is at least 3 GiB.
5.2. The disk customizations option
The disk customizations option provides a more powerful interface to control the whole partitioning layout of the image.
- Allowed mountpoints
When using
bootc-image-builder
, only the following directories allow customization:-
The
/
(root) directory. -
Custom directories under
/var
, but not/var
itself.
-
The
- Not allowed mountpoints
Under
/var
, the following mount points do not allow customization:-
/var/home
-
/var/lock
-
/var/mail
-
/var/mnt
-
/var/roothome
-
/var/run
-
/var/srv
-
/var/usrlocal
-
5.3. Describing disk customizations in a blueprint
When using the Disk customizations, you can describe the partition table almost entirely by using a blueprint. The customizations have the following structure:
Partitions: The top level is a list of partitions.
type
: Each partition has a type, which can be eitherplain
orlvm
. If the type is not set, it defaults toplain
. The remaining required and optional properties of the partition depend on the type.plain
: A plain partition is a partition with a filesystem. It supports the following properties:-
fs_type
: The filesystem type, which should be one ofxfs
,ext4
,vfat
, orswap
. Setting it toswap
will create a swap partition. The mountpoint for a swap partition must be empty. -
minsize
: The minimum size of the partition, as an integer (in bytes) or a string with a data unit (for example 3 GiB). The final size of the partition in the image might be larger for specific mountpoints. See Understanding partitions section. -
mountpoint
The mountpoint for the filesystem. For swap partitions, this must be empty. -
label
: The label for the filesystem (optional).
-
lvm
: An lvm partition is a partition with an LVM volume group. Only single Persistent Volumes volume groups are supported. It supports the following properties:-
name
: The name of the volume group (optional; if unset, a name will be generated automatically). -
minsize
: The minimum size of the volume group, as an integer (in bytes) or a string with a data unit (for example 3 GiB). The final size of the partition and volume group in the image might be larger if the value is smaller than the sum of logical volumes it contains. logical_volumes
: One or more logical volumes for the volume group. Each volume group supports the following properties:-
name
: The name of the logical volume (optional; if unset, a name will be generated automatically based on the mountpoint). -
minsize
: The minimum size of the logical volume, as an integer (in bytes) or a string with a data unit (for example 3 GiB). The final size of the logical volume in the image might be larger for specific mountpoints. See the General principles chapter for more details). -
label
: The label for the filesystem (optional). -
fs_type
: The filesystem type, which should be one ofxfs
,ext4
,vfat
, orswap
. Setting it to swap will create a swap logical volume. The mountpoint for a swap logical volume must be empty. -
mountpoint
: The mountpoint for the logical volume’s filesystem. For swap logical volumes, this must be empty.
-
-
- Order:
The order of each element in a list is respected when creating the partition table. The partitions are created in the order they are defined, regardless of their type.
- Incomplete partition tables:
Incomplete partitioning descriptions are valid. Partitions, LVM logical volumes, are added automatically to create a valid partition table. The following rules are applied:
-
A root filesystem is added if one is not defined. This is identified by the mountpoint
/
. If an LVM volume group is defined, the root filesystem is created as a logical volume, otherwise it will be created as a plain partition with a filesystem. The type of the filesystem, for plain and LVM, depends on the distribution (xfs for RHEL and CentOS, ext4 for Fedora). See Understanding partitions section for information about the sizing and order of the root partition. -
A boot partition is created if needed and if one is not defined. This is identified by the mountpoint
/boot
. A boot partition is needed when the root partition (mountpoint/
) is on an LVM logical volume. It is created as the first partition after the ESP (see next item). -
An EFI system partition (ESP) is created if needed. This is identified by the mountpoint
/boot/efi
. An ESP is needed when the image is configured to boot with UEFI. This is defined by the image definition and depends on the image type, the distribution, and the architecture. The type of the filesystem is alwaysvfat
. By default, the ESP is 200 MiB and is the first partition after the BIOS boot (see next item). - A 1 MiB unformatted BIOS boot partition is created at the start of the partition table if the image is configured to boot with BIOS. This is defined by the image definition and depends on the image type, the distribution, and the architecture. Both a BIOS boot partition and an ESP are created for images that are configured for hybrid boot.
- Combining partition types:
You can define multiple partitions. The following combination of partition types are valid:
-
plain
andlvm
: Plain partitions can be created alongside an LVM volume group. However, only one LVM volume group can be defined. - Examples: Blueprint to define two partitions
The following blueprint defines two partitions. The first is a 50 GiB partition with an ext4
filesystem that will be mounted at /data
. The second is an LVM volume group with three logical volumes, one for root /
, one for home directories /home
, and a swap
space in that order. The LVM volume group will have 15 GiB of non-allocated space.
[[customizations.disk.partitions]] type = "plain" label = "data" mountpoint = "/data" fs_type = "ext4" minsize = "50 GiB" [[customizations.disk.partitions]] type = "lvm" name = "mainvg" minsize = "20 GiB" [[customizations.disk.partitions.logical_volumes]] name = "rootlv" mountpoint = "/" label = "root" fs_type = "ext4" minsize = "2 GiB" [[customizations.disk.partitions.logical_volumes]] name = "homelv" mountpoint = "/home" label = "home" fs_type = "ext4" minsize = "2 GiB" [[customizations.disk.partitions.logical_volumes]] name = "swaplv" fs_type = "swap" minsize = "1 GiB"
5.4. The filesystem customization option
The filesystem customization option provides the final partition table of an image that you built with image builder, and it is determined by a combination of the following factors:
- The base partition table for a given image type.
The relevant blueprint customizations:
- Partitioning mode.
- Filesystem customizations.
The image size parameter of the build request:
-
On the command line, this is the
--size
option of thecomposer-cli compose start
command.
-
On the command line, this is the
The following describes how these factors affect the final layout of the partition table.
- Modifying partition tables
You can modify the partition table by taking the following aspects in consideration:
- Partitioning modes
The partitioning mode controls how the partition table is modified from the image type’s default layout.
-
The
raw
partition type does not convert any partition to LVM. -
The
lvm
partition type always converts the partition that contains the/
root mountpoint to an LVM Volume Group and creates a root Logical Volume. Except from/boot
, any extra mountpoint is added to the Volume Group as new Logical Volumes. The
auto-lvm
mode is the default mode and converts a raw partition table to an LVM-based one if and only if new mountpoints are defined in the filesystems customization. See the Mountpoints entry for more details.- Mountpoints
You can define new filesystems and minimum partition sizes by using the filesystems customization in the blueprint. By default, if new mountpoints are created, a partition table is automatically converted to LVM. See the Partitioning modes entry for more details.
- Image size The minimum size of the partition table is the size of the disk image. The final size of the image will either be the value of the size parameter or the sum of all partitions and their associated metadata, depending on which one is the larger.
5.5. Creating images with specific sizes
To create a disk image of a very specific size, you must ensure that the following requirements are met:
-
You must specify the exact
[Image size]
in the build request. - The mountpoints that you define as customizations must specify sizes that are smaller than the total size in sum. This is required, because the partition table, partitions, and other entities often require extra space for metadata and headers, so the space required to fit all the mountpoints is always larger than the sum of the size of the partitions. However, the exact size of the extra space that is required varies based on many factors, such as image type, for example.
The following are overall steps to create a disk image of a very specific size in the TOML file:
-
Set the
[Image size]
parameter to the size that you want. -
Add any extra mountpoints with their required minimum sizes. Ensure that the sum of the sizes is smaller than the image size by at least 3.01 GiB if there is no
/usr
mountpoint, or at least 1.01 GiB if there is. The extra 0.01 MiB is more than enough for the headers and metadata for which extra space might be reserved. -
Do not specify a size for the
/
mountpoint.
With this, you create a disk with a partition table of the desired size with each partition sized to fit the desired mountpoints. The root partition, root LVM Logical Volume, will be at least 3 GiB, or 1 GiB if /usr
is specified. See Understanding partitions for more details.
- If the partition table does not have any LVM Volume Groups (VG), the root partition will be grown to fill the remaining space.
- If the partition table contains an LVM Volume Group (VG), the VG will have unallocated extents that can be used to grow any of the Logical Volumes.
5.6. Building disk images of image-mode RHEL with advanced partitioning
Create image mode disk images with advanced partitioning by using bootc-image-builder
. The image mode disk images that you create of image mode RHEL with custom mount points, include custom mount options, LVM-based partitions and LVM-based SWAP. With that you can, for example, change the size of the /
and the /boot
directories by using a config.toml
file. When installing the RHEL image mode on bare-metal machines, you can benefit from all partitioning features available on Anaconda.
Prerequisites
- You have Podman installed on your host machine.
-
You have
virt-install
installed on your host machine. -
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Create a
config.toml
to configure custom mount options, for example:name = "lvm" description = "A base system with custom LVM partitioning" [customizations.disk] [[customizations.disk.partitions]] mountpoint = "/var/data" minsize = "1 GiB" label = "data" fs_type = "ext4" [[customizations.disk.partitions]] mountpoint = "/" minsize = "2 GiB" label = "root" fs_type = "ext4"
Run
bootc-image-builder
, passing theconfig.toml
as an argument.NoteIf you do not have the container storage mount and --local image options, your image must be public.
$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v ./config.toml:/config.toml \ -v ./output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type <image_type>\ --config config.toml \ quay.io/<namespace>/<image>:<tag>
You can find your image with the customized advanced partitioning in the output folder.
Next steps
- Deploy the disk image with advanced partitioning layout. See Deploying your customized images.
Additional resources
Chapter 6. Best practices for running containers using local sources
You can access content hosted in an internal registry that requires a custom Transport Layer Security (TLS) root certificate, when running RHEL bootc images.
There are two options available to install content to a container by using only local resources:
-
Bind mounts: Use for example
-v /etc/pki:/etc/pki
to override the container’s store with the host’s. -
Derived image: Create a new container image with your custom certificates by building it using a
Containerfile
.
You can use the same techniques to run a bootc-image-builder`
container or a bootc
container when appropriate.
6.1. Importing custom certificate to a container by using bind mounts
Use bound mounts to override the container’s store with the host’s.
Procedure
Run RHEL bootc image and use bind mount, for example
-v /etc/pki:/etc/pki
, to override the container’s store with the host’s:# podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/output:/output \ -v /etc/pki:/etc/pki \ localhost/<image> \ --type iso \ --config /config.toml \ quay.io/<namespace>/<image>:<tag>
Verification
List certificates inside the container:
# ls -l /etc/pki
6.2. Importing custom certificates to a container by using Containerfile
Create a new container image with your custom certificates by building it using a Containerfile
.
Procedure
Create a
Containerfile
:FROM <internal_repository>/<image> RUN mkdir -p /etc/pki/ca-trust/extracted/pem/ COPY tls-ca-bundle.pem /etc/pki/ca-trust/extracted/pem/ RUN rm -rf /etc/yum.repos.d/* COPY echo-rhel9_4.repo /etc/yum.repos.d/
Build the custom image:
# podman build -t <your_image> .
Run the
<your_image>
:# podman run -it --rm <your_image>
Verification
List the certificates inside the container:
# ls -l /etc/pki/ca-trust/extracted/pem/ tls-ca-bundle.pem
Chapter 7. Deploying the RHEL bootc images
You can deploy the rhel-bootc
container image by using the following different mechanisms.
- Anaconda
-
bootc-image-builder
-
bootc install
The following bootc image types are available:
Disk images that you generated by using the
bootc image-builder
such as:- QCOW2 (QEMU copy-on-write, virtual disk)
- Raw (Mac Format)
- AMI (Amazon Cloud)
- ISO: Unattended installation method, by using an USB Sticks or Install-on-boot.
After you have created a layered image that you can deploy, there are several ways that the image can be installed to a host:
You can use RHEL installer and Kickstart to install the layered image to a bare metal system, by using the following mechanisms:
- Deploy by using USB
- PXE
-
You can also use
bootc-image-builder
to convert the container image to a bootc image and deploy it to a bare metal or to a cloud environment.
The installation method happens only one time. After you deploy your image, any future updates will apply directly from the container registry as the updates are published.
Figure 7.1. Deploying a bootc image by using a basic build installer bootc install
, or deploying a container image by using Anaconda and Kickstart

Figure 7.2. Using bootc-image-builder
to create disk images from bootc images and deploying disk images in different environments, such as the edge, servers, and clouds by using Anaconda, bootc-image-builder
or bootc install

7.1. Deploying a container image by using KVM with a QCOW2 disk image
After creating a QEMU disk image from a RHEL bootc image by using the bootc-image-builder
tool, you can use a virtualization software to boot it.
Prerequisites
- You created a container image. See Creating QCOW2 images by using bootc-image-builder.
- You pushed the container image to an accessible repository.
Procedure
Run the container image that you create by using either
libvirt
. See Creating virtual machines by using the command line for more details.The following example uses
libvirt
:$ sudo virt-install \ --name bootc \ --memory 4096 \ --vcpus 2 \ --disk qcow2/disk.qcow2 \ --import \ --os-variant rhel9-unknown
Verification
- Connect to the VM in which you are running the container image. See Connecting to virtual machines for more details.
Next steps
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
Additional resources
7.2. Deploying a container image to AWS with an AMI disk image
After using the bootc-image-builder
tool to create an AMI from a bootc image, and uploading it to a AWS s3 bucket, you can deploy a container image to AWS with the AMI disk image.
Prerequisites
- You created an Amazon Machine Image (AMI) from a bootc image. See Creating AMI images by using bootc-image-builder and uploading it to AWS.
-
cloud-init
is available in the Containerfile that you previously created so that you can create a layered image for your use case.
Procedure
- In a browser, access Service→EC2 and log in.
- On the AWS console dashboard menu, choose the correct region. The image must have the Available status, to indicate that it was correctly uploaded.
- On the AWS dashboard, select your image and click .
- In the new window that opens, choose an instance type according to the resources you need to start your image. Click .
- Review your instance details. You can edit each section if you need to make any changes. Click .
- Before you start the instance, select a public key to access it. You can either use the key pair you already have or you can create a new key pair.
Click Initializing.
to start your instance. You can check the status of the instance, which displays asAfter the instance status is Running, the button becomes available.
- Click . A window appears with instructions on how to connect by using SSH.
Run the following command to set the permissions of your private key file so that only you can read it. See Connect to your Linux instance.
$ chmod 400 <your-instance-name.pem>
Connect to your instance by using its Public DNS:
$ ssh -i <your-instance-name.pem>ec2-user@<your-instance-IP-address>
Your instance continues to run unless you stop it.
Verification
After launching your image, you can:
- Try to connect to http://<your_instance_ip_address> in a browser.
- Check if you are able to perform any action while connected to your instance by using SSH.
Next steps
- After you deploy your image, you can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
Additional resources
7.3. Deploying a container image by using Anaconda and Kickstart
You can deploy the RHEL ISO image that you downloaded from Red Hat by using Anaconda and Kickstart to install your container image.
The use of rpm-ostree
to make changes, or install content, is not supported.
Prerequisites
- You have downloaded the 9.4 Boot ISO for your architecture from Red Hat. See Downloading RH boot images.
Procedure
Create an
ostreecontainer
Kickstart file. For example:# Basic setup text network --bootproto=dhcp --device=link --activate # Basic partitioning clearpart --all --initlabel --disklabel=gpt reqpart --add-boot part / --grow --fstype xfs # Reference the container image to install - The kickstart # has no
%packages
section. A container image is being installed. ostreecontainer --url registry.redhat.io/rhel9/rhel-bootc:9.4 firewall --disabled services --enabled=sshd # Only inject a SSH key for root rootpw --iscrypted locked sshkey --username root "<your key here>" rebootBoot a system by using the 9.4 Boot ISO installation media.
Append the Kickstart file with the following to the kernel argument:
inst.ks=http://<path_to_your_kickstart>
- Press CTRL+X to boot the system.
Next steps
- After you deploy your container image, you can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
Additional resources
- ostreecontainer documentation
- bootc upgrade fails when using local rpm-ostree modifications (Red Hat Knowledgebase)
7.4. Deploying a custom ISO container image
After you build an ISO image by using bootc-image-builder
, the resulting image is a system similar to the RHEL ISOs available for download, except that your container image content is embedded in the ISO disk image. You do not need to have access to the network during installation. You can install the resulting ISO disk image to a bare metal system. See Creating ISO images by using bootc-image-builder.
Prerequisites
- You have created an ISO image with your bootc image embedded.
Procedure
- Copy your ISO disk image to a USB flash drive.
- Perform a bare-metal installation by using the content in the USB stick into a disconnected environment.
Next steps
- After you deploy your container image, you can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
7.5. Deploying an ISO bootc image over PXE boot
You can use a network installation to deploy the RHEL ISO image over PXE boot to run your ISO bootc image.
Prerequisites
- You have downloaded the 9.4 Boot ISO for your architecture from Red Hat. See Downloading RH boot images.
You have configured the server for the PXE boot. Choose one of the following options:
- For HTTP clients, see Configuring the DHCPv4 server for HTTP and PXE boot.
- For UEFI-based clients, see Configuring a TFTP server for UEFI-based clients.
- For BIOS-based clients, see Configuring a TFTP server for BIOS-based clients.
- You have a client, also known as the system to which you are installing your ISO image.
Procedure
- Export the RHEL installation ISO image to the HTTP server. The PXE boot server is now ready to serve PXE clients.
- Boot the client and start the installation.
- Select PXE Boot when prompted to specify a boot source. If the boot options are not displayed, press the Enter key on your keyboard or wait until the boot window opens.
- From the Red Hat Enterprise Linux boot window, select the boot option that you want, and press Enter.
- Start the network installation.
Next steps
- You can make updates to the image and push the changes to a registry. See Managing RHEL bootc images.
7.6. Building, configuring, and launching disk images with bootc-image-builder
You can inject configuration into a custom image by using a Containerfile.
Procedure
Create a disk image. The following example shows how to add a user to the disk image.
[[blueprint.customizations.user]] name = "user" password = "pass" key = "ssh-rsa AAA ... user@email.com" groups = ["wheel"]
-
name
- User name. Mandatory -
password
- Nonencrypted password. Not mandatory -
key
- Public SSH key contents. Not mandatory -
groups
- An array of groups to add the user into. Not mandatory
-
Run
bootc-image-builder
and pass the following arguments:$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml \ -v $(pwd)/output:/output \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --type qcow2 \ --config config.toml \ quay.io/<namespace>/<image>:<tag>
Launch a VM, for example, by using
virt-install
:$ sudo virt-install \ --name bootc \ --memory 4096 \ --vcpus 2 \ --disk qcow2/disk.qcow2 \ --import \ --os-variant rhel9
Verification
Access the system with SSH:
# ssh -i /<path_to_private_ssh-key> <user1>@<ip-address>
Next steps
- After you deploy your container image, you can make updates to the image and push the changes to a registry. See Managing RHEL bootable images.
7.7. Deploying a container image by using bootc
With bootc
, you have a container that is the source of truth. It contains a basic build installer and it is available as bootc install to-disk
or bootc install to-filesystem
. By using the bootc install
methods you do not need to perform any additional steps to deploy the container image, because the container images include a basic installer.
With image mode for RHEL, you can install unconfigured images, for example, images that do not have a default password or SSH key.
Perform a bare-metal installation to a device by using a RHEL ISO image.
Prerequisites
- You have downloaded the 9.4 Boot ISO for your architecture from Red Hat. See Downloading RH boot images.
- You have created a configuration file.
Procedure
inject a configuration into the running ISO image, for example:
$ podman run --rm --privileged --pid=host -v /var/lib/containers:/var/lib/containers --security-opt label=type:unconfined_t <image> bootc install to-disk <path-to-disk>
Next steps
- After you deploy your container image, you can make updates to the image and push the changes to a registry. See Managing RHEL bootable images.
7.8. Advanced installation with to-filesystem
The bootc install
contains two subcommands: bootc install to-disk
and bootc install to-filesystem
.
-
The
bootc-install-to-filesystem
performs installation to the target filesystem. The
bootc install to-disk
subcommand consists of a set of opinionated lower level tools that you can also call independently. The command consist of the following tools:-
mkfs.$fs /dev/disk
-
mount /dev/disk /mnt
-
bootc install to-filesystem --karg=root=UUID=<uuid of /mnt> --imgref $self /mnt
-
7.8.1. Using bootc install to-existing-root
The bootc install to-existing-root
is a variant of install to-filesystem
. You can use it to convert an existing system into the target container image.
This conversion eliminates the /boot
and /boot/efi
partitions and can delete the existing Linux installation. The conversion process reuses the filesystem, and even though the user data is preserved, the system no longer boots in package mode.
Prerequisites
- You must have root permissions to complete the procedure.
-
You must match the host environment and the target container version, for example, if your host is a RHEL 9 host, then you must have a RHEL 9 container. Installing a RHEL container on a Fedora host by using
btrfs
as the RHEL kernel will not support that filesystem.
Procedure
Run the following command to convert an existing system into the target container image. Pass the target
rootfs
by using the-v /:/target
option.# podman run --rm --privileged -v /dev:/dev -v /var/lib/containers:/var/lib/containers -v /:/target \ --pid=host --security-opt label=type:unconfined_t \ <image> \ bootc install to-existing-root
This command deletes the data in
/boot
, but everything else in the existing operating system is not automatically deleted. This can be useful because the new image can automatically import data from the previous host system. Consequently, container images, database, the user home directory data, configuration files in/etc
are all available after the subsequent reboot in/sysroot
.You can also use the
--root-ssh-authorized-keys
flag to inherit the root user SSH keys, by adding--root-ssh-authorized-keys /target/root/.ssh/authorized_keys
. For example:# podman run --rm --privileged -v /dev:/dev -v /var/lib/containers:/var/lib/containers -v /:/target \ --pid=host --security-opt label=type:unconfined_t \ <image> \ bootc install to-existing-root --root-ssh-authorized-keys /target/root/.ssh/authorized_keys
Chapter 8. Enabling the FIPS mode while building a bootc image
FIPs include standards for cryptographic operations. You can enable the FIPS mode during the bootc image build time, when building a bootc image, to configure the system to use only FIPS approved modules. There are 2 options to enable FIPS mode:
-
By using the
bootc-image-builder
tool: you must enable the FIPS crypto policy into the Containerfile. -
When performing an Anaconda installation: apart from enabling the FIPS crypto policy into the Containerfile, you must add the
fips=1
kernel argument during the boot time.
FIPS dracut module is built-in to the base image. It defaults to a boot=UUID= karg
in bootc install-to-filesystem
.
8.1. Enabling the FIPS mode by using bootc-image-builder
Create a disk image by using bootc-image-builder
or bootc install to-disk
, and enable the FIPS mode by passing the custom Containerfile as an argument when building the image.
Prerequisites
- You have Podman installed on your host machine.
-
You have
virt-install
installed on your host machine. -
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Create a
01-fips.toml
to configure FIPS enablement, for example:# Enable FIPS kargs = ["fips=1"]
Create a Containerfile with the following instructions to enable the
fips=1
kernel argument:FROM registry.redhat.io/rhel9/rhel-bootc:latest # Enable fips=1 kernel argument: https://containers.github.io/bootc/building/kernel-arguments.html COPY 01-fips.toml /usr/lib/bootc/kargs.d/ # Enable the FIPS crypto policy # crypto-policies-scripts is not installed by default in RHEL-10 RUN dnf install -y crypto-policies-scripts && update-crypto-policies --no-reload --set FIPS
Create your bootc
<image>
compatible base disk image by usingContainerfile
in the current directory:$ podman build -t quay.io/<namespace>/<image>:<tag> .
Verification
After login in to the system, check that FIPS mode is enabled:
$ *fips-mode-setup --check* FIPS mode is enabled.
Additional resources
8.2. Enabling the FIPS mode to perform an Anaconda installation
To create a disk image and enable the FIPS mode when performing an Anaconda installation, follow the steps:
Prerequisites
- You have Podman installed on your host machine.
-
You have
virt-install
installed on your host machine. -
You have root access to run the
bootc-image-builder
tool, and run the containers in--privileged
mode, to build the images.
Procedure
Create a
01-fips.toml
to configure FIPS enablement, for example:# Enable FIPS kargs = ["fips=1"]
Create a Containerfile with the following instructions to enable the
fips=1
kernel argument:FROM registry.redhat.io/rhel9/rhel-bootc:latest # Enable fips=1 kernel argument: https://containers.github.io/bootc/building/kernel-arguments.html COPY 01-fips.toml /usr/lib/bootc/kargs.d/ # Install and enable the FIPS crypto policy RUN dnf install -y crypto-policies-scripts && update-crypto-policies --no-reload --set FIPS
Create your bootc
<image>
compatible base disk image by usingContainerfile
in the current directory:$ sudo podman run \ --rm \ -it \ --privileged \ --pull=newer \ --security-opt label=type:unconfined_t \ -v $(pwd)/config.toml:/config.toml:ro \ -v $(pwd)/output:/output \ -v /var/lib/containers/storage:/var/lib/containers/storage \ registry.redhat.io/rhel9/bootc-image-builder:latest \ --local --type iso \ quay.io/<namespace>/<image>:<tag>
Enable FIPS mode during the system installation:
When booting the RHEL Anaconda installer, on the installation screen, press the TAB key and add the
fips=1
kernel argument.After the installation, the system starts in FIPS mode automatically.
Verification
After login in to the system, check that FIPS mode is enabled:
$ *fips-mode-setup --check* FIPS mode is enabled.
Additional resources
Chapter 9. Managing RHEL bootc images
After installing and deploying RHEL bootc images, you can perform management operations on your container images, such as changing or updating the systems. The system supports in-place transactional updates with rollback after deployment.
This kind of management, also known as Day 2 management baseline, consists of transactionally fetching new operating system updates from a container registry and booting the system into them, while supporting manual, or automated rollbacks in case of failures.
You can also rely on automatic updates, that are turned on by default. The systemd service unit
and the systemd timer unit
files check the container registry for updates and apply them to the system. You can trigger an update process with different events, such as updating an application. There are automation tools watching these updates and then triggering the CI/CD pipelines. A reboot is required, because the updates are transactional. For environments that require more sophisticated or scheduled rollouts, you must disable auto updates and use the bootc
utility to update your operating system.
See Day 2 operations support for more details.
Figure 9.1. Manually updating an installed operating system, changing the container image reference or rolling back changes if needed

9.1. Switching the container image reference
You can change the container image reference used for upgrades by using the bootc switch
command. For example, you can switch from the stage to the production tag. The bootc switch
command performs the same operations as the bootc upgrade
command and additionally changes the container image reference.
To manually switch an existing ostree-based
container image reference, use the bootc switch
command.
The use of rpm-ostree
to make changes, or install content, is not supported.
Prerequisites
-
A booted system using
bootc
.
Procedure
Run the following command:
$ sudo bootc switch [--apply] quay.io/<namespace>/<image>:<tag>
Optionally, you can use the
--apply
option when you want to automatically take actions, such as rebooting if the system has changed.
The bootc switch
command has the same effect as bootc upgrade
. The only difference is the container image reference is changed. This allows preserving the existing states in /etc
and /var
, for example, host SSH keys and home directories.
Additional resources
-
The
bootc-switch
man page
9.2. Adding modules to the bootc image initramfs
The rhel9/rhel-bootc
image uses the dracut
infrastructure to build an initial RAM disk, the initrd
during the image build time. The initrd
is built and included in the /usr/lib/modules/$kver/initramfs.img
location inside the container.
You can use a drop-in configuration file to override the dracut
configuration, and place it in /usr/lib/dracut/dracut.conf.d/<50-custom-added-modules.conf>
And thus re-create initrd
with the modules you want to add.
Prerequisites
- A booted system using bootc.
Procedure
Re-create the
initrd
as part of a container build:FROM <baseimage> COPY <50-custom-added-modules>.conf /usr/lib/dracut/dracut.conf.d RUN set -x; kver=$(cd /usr/lib/modules && echo *); dracut -vf /usr/lib/modules/$kver/initramfs.img $kver
NoteBy default the command attempts to pull the running kernel version, which causes an error. Explicitly pass to
dracut
the kernel version of the target to avoid errors.
9.3. Modifying and regenerating initrd
The default container image includes a pre-generated initial RAM disk (initrd) in /usr/lib/modules/$kver/initramfs.img
. To regenerate the initrd
, for example, to add a dracut module, follow the steps:
Procedure
Write your drop-in configuration file. For example:
dracutmodules = "module"
Place your drop-in configuration file in the location that
dracut
normally uses:/usr
. For example:/usr/lib/dracut/dracut.conf.d/50-custom-added-modules.conf
Regenerate the
initrd
as part of the container build. You must explicitly pass the kernel version to target todracut
, because it tries to pull the running kernel version, which can cause an error. The following is an example:FROM <baseimage> COPY 50-custom-added-modules.conf /usr/lib/dracut/dracut.conf.d RUN set -x; kver=$(cd /usr/lib/modules && echo *); dracut -vf /usr/lib/modules/$kver/initramfs.img $kver
9.4. Performing manual updates from an installed operating system
Installing image mode for RHEL is a one time task. You can perform any other management task, such as changing or updating the system, by pushing the changes to the container registry.
When using image mode for RHEL, you can choose to perform manual updates for your systems. Manual updates are also useful if you have an automated way to perform updates, for example, by using Ansible. Because the automatic updates are enabled by default, to perform manual updates you must turn the automatic updates off. You can do this by choosing one of the following options:
-
Running the
bootc upgrade
command -
Modifying the
systemd
timer file
9.5. Turning off automatic updates
To perform manual updates you must turn off automatic updates. You can do this by choosing one of the following options in the procedure below.
Procedure
Disable the timer of a container build.
By running the
systemctl mask
command:$ systemctl mask bootc-fetch-apply-updates.timer
By modifying the
systemd
timer file. Usesystemd
"drop-ins" to override the timer. In the following example, updates are scheduled for once a week.Create an
updates.conf
file with the following content:[Timer] # Clear previous timers OnBootSec= OnBootSec=1w OnUnitInactiveSec=1w
Add you file to the directory you created:
$ mkdir -p /usr/lib/systemd/system/bootc-fetch-apply-updates.timer.d $ cp updates.conf /usr/lib/systemd/system/bootc-fetch-apply-updates.timer.d
9.6. Manually updating an installed operating system
To manually fetch updates from a registry and boot the system into the new updates, use bootc upgrade
. This command fetches the transactional in-place updates from the installed operating system to the container image registry. The command queries the registry and queues an updated container image for the next boot. It stages the changes to the base image, while not changing the running system by default.
Procedure
Run the following command:
$ bootc upgrade [--apply]
The
apply
argument is optional and you can use it when you want to automatically take actions, such as rebooting if the system has changed.
The bootc upgrade
and bootc update
commands are aliases.
Additional resources
-
The
bootc-upgrade
man page
9.7. Performing rollbacks from a updated operating system
You can roll back to a previous boot entry to revert changes by using the bootc rollback
command. This command changes the boot loader entry ordering by making the deployment under rollback
queued for the next boot. The current deployment then becomes the rollback. Any staged changes, such as a queued upgrade that was not applied, are discarded.
After a rollback completes, the system reboots and the update timer runs within 1 to 3 hours which automatically updates and reboots your system to the image you just rolled back from.
If you perform a rollback, the system will automatically update again unless you turn off auto-updates. See Turning off automatic updates.
Prerequisites
- You performed an update to the system.
Procedure
Run the following command:
$ bootc rollback [-h|--help] [-V|--version]
The bootc rollback
command has the same effect as bootc upgrade
. The only difference is the container image being tracked. This enables preserving the existing states in /etc
and /var
, for example, host SSH keys and home directories.
Verification
Use
systemd journal
to check the logged message for the detected rollback invocation.$ journalctl -b
You can see a log similar to:
MESSAGE_ID=26f3b1eb24464d12aa5e7b544a6b5468
Additional resources
-
The
bootc-rollback
man page
9.8. Deploying updates to system groups
You can change the configuration of your operating system by modifying the Containerfile. Then you can build and push your container image to the registry. When you next boot your operating system, an update will be applied.
You can also change the container image source by using the bootc switch
command. The container registry is the source of truth. See Switching the container image reference.
Usually, when deploying updates to system groups, you can use a central management service to provide a client to be installed on each system which connects to the central service. Often, the management service requires the client to perform a one time registration. The following is an example on how to deploy updates to system groups. You can modify it to create a persistent systemd
service, if required.
For clarity reasons, the Containerfile in the example is not optimized. For example, a better optimization to avoid creating multiple layers in the image is by invoking RUN a single time.
You can install a client into an image mode for RHEL image and run it at startup to register the system.
Prerequisites
-
The management-client handles future connections to the server, by using a
cron
job or a separatesystemd
service.
Procedure
Create a management service with the following characteristics. It determines when to upgrade the system.
-
Disable
bootc-fetch-apply-updates.timer
if it is included in the base image. -
Install the client by using
dnf
, or some other method that applies for your client. - Inject the credentials for the management service into the image.
-
Disable
9.9. Checking inventory health
Health checks are one of the Day 2 Operations. You can manually check the system health of the container images and events that are running inside the container.
You can set health checks by creating the container on the command line. You can display the health check status of a container by using the podman inspect
or podman ps
commands.
You can monitor and print events that occur in Podman by using the podman events
command. Each event includes a timestamp, a type, a status, a name, if applicable, and an image, if applicable.
For more information about health checks and events, see chapter Monitoring containers.
9.10. Automation and GitOps
You can automate the building process by using CI/CD pipelines so that an update process can be triggered by events, such as updating an application. You can use automation tools that track these updates and trigger the CI/CD pipelines. The pipeline keeps the systems up to date by using the transactional background operating system updates.
9.11. Using Toolbx to inspect bootc containers
Installing software on a system presents certain risks: it can change a system’s behavior, and can leave unwanted files and directories behind after they are no longer needed. You can prevent these risks by installing your favorite development and debugging tools, editors, and software development kits (SDKs) into the Toolbx utility included in the RHEL bootc, an image fully mutable container without affecting the base operating system. You can perform changes on the host system with commands such as less
, lsof
, rsync
, ssh
, sudo
, and unzip
.
The Toolbx utility performs the following actions:
-
Pulling the
registry.access.redhat.com/ubi9/toolbox:latest
image to your local system - Starting up a container from the image
- Running a shell inside the container from which you can access the host system
Toolbx can run a root container or a rootless container, depending on the rights of the user who creates the Toolbx container. Utilities that would require root rights on the host system also should be run in root containers.
The default container name is rhel-toolbox
. To inspect bootc containers, follow the steps:
Procedure
Start a Toolbx container by using the
toolbox create
command and enter the container with thetoolbox enter
command.As a rootless user:
$ toolbox create <mytoolbox>
As a root user:
$ sudo toolbox create <mytoolbox> Created container: <mytoolbox> Enter with: toolbox enter
Verify that you pulled the correct image:
[user@toolbox ~]$ toolbox list IMAGE ID IMAGE NAME CREATED fe0ae375f149 registry.access.redhat.com/ubi{ProductVersion}/toolbox 5 weeks ago CONTAINER ID CONTAINER NAME CREATED STATUS IMAGE NAME 5245b924c2cb <mytoolbox> 7 minutes ago created registry.access.redhat.com/ubi{ProductVersion}/toolbox:8.9-6
Enter the Toolbx container:
[user@toolbox ~]$ toolbox enter <mytoolbox>
- Optional: Check if you pulled the correct image
Enter a command inside the
<mytoolbox>
container and display the name of the container and the image:⬢ [user@toolbox ~]$ cat /run/.containerenv engine="podman-4.8.2" name="<mytoolbox>" id="5245b924c2cb..." image="registry.access.redhat.com/ubi{ProductVersion}/toolbox" imageid="fe0ae375f14919cbc0596142e3aff22a70973a36e5a165c75a86ea7ec5d8d65c"
Use the Toolbx to install the development tools:
Install the tools of your choice, for example, the Emacs text editor, GCC compiler and GNU Debugger (GDB):
⬢[user@toolbox ~]$ sudo dnf install emacs gcc gdb
Optional: Verify that the tools are installed:
⬢[user@toolbox ~]$ dnf repoquery --info --installed <package_name>
After installation, you can continue using those tools as a rootless user.
Use Toolbx to troubleshoot the host system without installing them on the host system.
Install the
systemd
suite to be able to run thejournalctl
command:⬢[root@toolbox ~]# dnf install systemd
Display log messages for all processes running on the host:
⬢[root@toolbox ~]# j journalctl --boot -0 Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: microcode: updated ear> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: Linux version 6.6.8-10> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: Command line: BOOT_IMA> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: x86/split lock detecti> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: BIOS-provided physical>
Display log messages for the kernel:
⬢[root@toolbox ~]# journalctl --boot -0 --dmesg Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: microcode: updated ear> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: Linux version 6.6.8-10> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: Command line: BOOT_IMA> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: x86/split lock detecti> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: BIOS-provided physical> Jan 02 09:06:48 user-thinkpadp1gen4i.brq.csb kernel: BIOS-e820: [mem 0x0000>
Install the
nmap
network scanning tool:⬢[root@toolbox ~]# dnf install nmap
Scan IP addresses and ports in a network:
⬢[root@toolbox ~]# nmap -sS scanme.nmap.org Starting Nmap 7.93 ( https://nmap.org ) at 2024-01-02 10:39 CET Stats: 0:01:01 elapsed; 0 hosts completed (0 up), 256 undergoing Ping Scan Ping Scan Timing: About 29.79% done; ETC: 10:43 (0:02:24 remaining) Nmap done: 256 IP addresses (0 hosts up) scanned in 206.45 seconds
-
The
-sS
option performs a TCP SYN scan. Most of Nmap’s scan types are only available to privileged users, because they send and receive raw packets, which requires root access on UNIX systems.
-
The
Stop the Toolbx bootc container.
Leave the container and return to the host:
⬢ [user@toolbox ~]$ exit
Stop the toolbox container:
⬢ [user@toolbox ~]$ podman stop <mytoolbox>
Optional: Remove the toolbox container:
⬢ [user@toolbox ~]$ toolbox rm <mytoolbox>
Alternatively, you can also use the
podman rm
command to remove the bootc container.
Additional resources
- Debugging image mode hosts article
Chapter 10. Managing kernel arguments in bootc systems
You can use bootc
to configure kernel arguments. By default, bootc
uses the boot loader configuration files that are stored in /boot/loader/entries
. This directory defines arguments provided to the Linux kernel. The set of kernel arguments is machine-specific state, but you can also manage the kernel arguments by using container updates. The boot loader menu entries are shared between multiple operating systems and boot loaders are installed on one device.
Currently, the boot loader entries are written by an OSTree backend.
10.1. How to add support to inject kernel arguments with bootc
The bootc
tool uses generic operating system kernels. You can add support to inject kernel arguments by adding a custom configuration, in the TOML format, in /usr/lib/bootc/kargs.d
. For example:
# /usr/lib/bootc/kargs.d/10-example.toml kargs = ["mitigations=auto,nosmt"]
You can also make these kernel arguments architecture-specific by using the match-architectures
key. For example:
# /usr/lib/bootc/kargs.d/00-console.toml kargs = ["console=ttyS0,114800n8"] match-architectures = ["x86_64"]
10.2. How to modify kernel arguments by using bootc install configs
You can use bootc install
to add kernel arguments during the install time in the following ways:
- Adding kernel arguments into the container image.
-
Adding kernel arguments by using the
bootc install --karg
command.
You can use the kernel arguments on Day 2 operations, by adding the arguments and applying them on a switch, upgrade, or edit. Adding kernel arguments and using it for Day 2 operations involves the following high-level steps:
-
Create files within
/usr/lib/bootc/kargs.d
with kernel arguments. - Fetch the container image to get the OSTree commit.
- Use the OSTree commit to return the file tree.
-
Navigate to
/usr/lib/bootc/kargs.d
. - Read each file within the directory.
-
Push the contents of each
kargs
file into a file containing all the neededkargs
. -
Pass the
kargs
to thestage()
function. - Apply these arguments to switch, upgrade, or edit.
10.3. How to inject kernel arguments in the Containerfile
To add kernel arguments into a container image, use a Containerfile. The following is an example:
FROM registry.redhat.io/rhel9/rhel-bootc:latest RUN mkdir -p /usr/lib/bootc/kargs.d RUN cat <<EOF >> /usr/lib/bootc/kargs.d/console.toml kargs = ["console=ttyS0,114800n8"] match-architectures = ["x86_64"] EOF RUN cat <<EOF >> /usr/lib/bootc/kargs.d/01-mitigations.toml kargs = ["mitigations=on", "systemd.unified_cgroup_hierarchy=0"] match-architectures = ["x86_64", "aarch64"] EOF
10.4. How to inject kernel arguments at installation time
You can use boot install
with the --karg
to inject kernel arguments during installation time. As a result, the kernel arguments become machine-local state.
For example, to inject kernel arguments, use the following command:
# bootc install to-filesystem --karg
Currently, bootc does not have an API to manipulate kernel arguments. This is only supported by rpm-ostree
, by using the rpm-ostree kargs
command.
10.5. How to add install-time kernel arguments with bootc-image-builder
The bootc-image-builder
tool supports the customizations.kernel.append
during install-time.
To add the kernel arguments with bootc-image-builder
, use the following customization:
{ "customizations": { "kernel": { "append": "mitigations=auto,nosmt" } } }
10.6. About changing kernel arguments post-install with kargs.d
The changes that you make to kargs.d
files and include in a container build are applied after the installation, and the difference between the set of kernel arguments is applied to the current boot loader configuration. This preserves any machine-local kernel arguments. You can use any tool to edit the /boot/loader/entries
files, which are in a standardized format. The /boot
file has read-only access to limit the set of tools that can write to this filesystem.
10.7. How to edit kernel arguments in bootc systems
To perform machine local changes, you also can edit kernel arguments on a bootc system or an`rpm-ostree` system, by using the rpm-ostree kargs
command. The changes are made through the user/lib/bootc/kargs.d
path, which also handles "Day 2" changes, besides the first boot changes.
The following are the options that you can use to add, modify or remove kernel arguments.
rpm-ostree kargs [option]
- --append=KEY=VALUE
-
Appends a kernel argument. It is useful with, for example,
console=
that can be used multiple times. You can use an empty value for an argument. - --replace=KEY=VALUE=NEWVALUE
-
Replaces an existing kernel argument. You can replace an argument with
KEY=VALUE
only if one value already exists for that argument. - --delete=KEY=VALUE
- Deletes a specific kernel key-value pair argument or an entire argument with a single key-value pair.
- --append-if-missing=KEY=VALUE
- Appends a kernel argument. Does nothing if the key is already present.
- --delete-if-present=KEY=VALUE
- Deletes a specific kernel key-value pair argument. Does nothing if the key is missing.
- --editor
- Uses an editor to modify the kernel arguments.
For more information, check the help:
# rpm-ostree kargs --help
The following is an example:
# rpm-ostree kargs --append debug Staging deployment... done Freed: 40.1 MB (pkgcache branches: 0) Changes queued for next boot. Run "systemctl reboot" to start a reboot
Chapter 11. Managing file systems in image mode for RHEL
Currently, image mode for RHEL uses OSTree as a backend, and enables composefs
for storage by default. The /opt
and /usr/local
paths are plain directories, and not symbolic links into /var
. This enables you to easily install third-party content in derived container images that write into /opt
for example.
11.1. Physical and logical root with /sysroot
When a system is fully booted, it is similar to chroot
, that is, the operating system changes the apparent root directory for the current running process and its children. The physical host root filesystem is mounted at /sysroot
. The chroot
filesystem is called a deployment root.
The remaining filesystem paths are part of a deployment root which is used as a final target for the system boot. The system uses the ostree=kernel
argument to find the deployment root.
/usr
-
This filesystem keeps all operating system content in
/usr
, with directories such as/bin
working as symbolic links to/usr/bin
.
composefs
enabled /usr
is not different from /
. Both directories are part of the same immutable image, so you do not need to perform a full UsrMove
with a bootc system.
/usr/local
-
The base image is configured with
/usr/local
as the default directory. /etc
The
/etc
directory contains mutable persistent state by default, but it supports enabling theetc.transient config
option. When the directory is in mutable persistent state, it performs a 3-way merge across upgrades:-
Uses the new default
/etc
as a base -
Applies the diff between current and previous
/etc
to the new/etc
directory -
Retains locally modified files that are different from the default
/usr/etc
of the same deployment in/etc
.
-
Uses the new default
The ostree-finalize-staged.service
executes these tasks during shutdown time, before creating the new boot loader entry.
This happens because many components of a Linux system ship default configuration files in the /etc
directory. Even if the default package does not ship it, by default the software only checks for config files in /etc
. Non bootc image based update systems with no distinct versions of /etc
are populated only during the installation time, and will not be changed at any point after installation. This causes the /etc
system state to be influenced by the initial image version and can lead to problems to apply a change, for example, to /etc/sudoers.conf
, and requires external intervention. For more details about file configuration, see Building and testing RHEL bootc images.
/var
-
The content in the
/var
directory is persistent by default. You can also make/var
or subdirectories mount points be persistent, whether network ortmpfs
.
There is just one /var
directory. If it is not a distinct partition, then physically the /var
directory is a bind mount into /ostree/deploy/$stateroot/var
and is shared across the available boot loader entries deployments.
By default, the content in /var
acts as a volume, that is, the content from the container image is copied during the initial installation time, and is not updated thereafter.
The /var
and the /etc
directories are different. You can use /etc
for relatively small configuration files, and the expected configuration files are often bound to the operating system binaries in /usr
. The /var
directory has arbitrarily large data, for example, system logs, databases, and by default, will not be rolled back if the operating system state is rolled back.
For example, making an update such as dnf downgrade postgresql
should not affect the physical database in /var/lib/postgres
. Similarly, making a bootc update
or bootc rollback
do not affect this application data.
Having /var
separate also makes it work cleanly to stage new operating system updates before applying them, that is, updates are downloaded and ready, but only take effect on reboot. The same applies for Docker volume, as it decouples the application code from its data.
You can use this case if you want applications to have a pre-created directory structure, for example, /var/lib/postgresql
. Use systemd tmpfiles.d
for this. You can also use StateDirectory=<directory>
in units.
- Other directories
-
There is no support to ship content in
/run
,/proc
or other API Filesystems in container images. Apart from that, other top level directories such as/usr
, and/opt
, are lifecycled with the container image. /opt
-
With
bootc
usingcomposefs
, the/opt
directory is read-only, alongside other top level directories such as/usr
.
When a software needs to write to its own directory in /opt/exampleapp
, a common pattern is to use a symbolic link to redirect to, for example, /var
for operations such as log files:
RUN rmdir /opt/exampleapp/logs && ln -sr /var/log/exampleapp /opt/exampleapp/logs
Optionally, you can configure the systemd unit to launch the service to do these mounts dynamically. For example:
BindPaths=/var/log/exampleapp:/opt/exampleapp/logs
- Enabling transient root
-
To enable a software to transiently (until the next reboot) write to all top-level directories, including
/usr
and/opt
, with symlinks to/var
for content that should persist, you can enable transient root. To enable a fully transient writablerootfs
by default, set the following option in/usr/lib/ostree/prepare-root.conf
.
[root] transient = true
This enables a software to transiently write to /opt
, with symlinks to /var
for content that must persist.
Additional resources
- Enabling transient root documentation
11.2. Version selection and bootup
Image mode for RHEL uses GRUB by default, with exception to s390x
architectures. Each version of image mode for RHEL currently available on a system has a menu entry.
The menu entry references an OSTree deployment which consists of a Linux kernel, an initramfs
and a hash linking to an OSTree commit, that you can pass by using the ostree=kernel
argument.
During bootup, OSTree reads the kernel argument to determine which deployment to use as the root filesystem. Each update or change to the system, such as package installation, addition of kernel arguments, creates a new deployment.
This enables rolling back to a previous deployment if the update causes problems.
Chapter 12. Appendix: Managing users, groups, SSH keys, and secrets in image mode for RHEL
Learn more about users, groups, SSH keys, and secrets management in image mode for RHEL.
12.1. Users and groups configuration
Image mode for RHEL is a generic operating system update and configuration mechanism. You cannot use it to configure users or groups. The only exception is the bootc install
command that has the --root-ssh-authorized-keys
option.
- Users and groups configuration for generic base images
- Usually, the distribution base images do not have any configuration. Do not encrypt passwords and SSH keys with publicly-available private keys in generic images because of security risks.
- Injecting SSH keys through
systemd
credentials -
You can use
systemd
to inject a root password or SSHauthorized_keys
file in some environments. For example, use System Management BIOS (SMBIOS) to inject SSH keys system firmware. You can configure this in local virtualization environments, such asqemu
. - Injecting users and SSH keys by using
cloud-init
-
Many Infrastructure as a service (IaaS) and virtualization systems use metadata servers that are commonly processed by software such as
cloud-init
orignition
. See AWS instance metadata. The base image you are using might includecloud-init
or Ignition, or you can install it in your own derived images. In this model, the SSH configuration is managed outside of the bootc image. - Adding users and credentials by using container or unit custom logic
-
Systems such as
cloud-init
are not privileged. You can inject any logic you want to manage credentials in the way you want to launch a container image, for example, by using asystemd
unit. To manage the credentials, you can use a custom network-hosted source, for example, FreeIPA. - Adding users and credentials statically in the container build
In package-oriented systems, you can use the derived build to inject users and credentials by using the following command:
RUN useradd someuser
You can find issues in the default
shadow-utils
implementation ofuseradd
: Users and groups IDs are allocated dynamically, and this can cause drift.- User and group home directories and
/var
directory For systems configured with persistent
/home → /var/home
, any changes to/var
made in the container image after initial installation will not be applied on subsequent updates.For example, if you inject
/var/home/someuser/.ssh/authorized_keys
into a container build, existing systems do not get the updatedauthorized_keys
file.- Using DynamicUser=yes for
systemd
units Use the
systemd
DynamicUser=yes
option where possible for system users.This is significantly better than the pattern of allocating users or groups at package install time, because it avoids potential UID or GID drift.
- Using
systemd
-sysusers Use
systemd
-sysusers, for example, in your derived build. For more information, see thesystemd
-sysusers documentation.COPY mycustom-user.conf /usr/lib/sysusers.d
The
sysusers
tool makes changes to the traditional/etc/passwd
file as necessary during boot time. If/etc
is persistent, this can avoidUID
orGID
drift. It means that theUID
orGID
allocation depends on how a specific machine was upgraded over time.- Using
systemd
JSON user records -
See JSON user records
systemd
documentation. Unlikesysusers
, the canonical state for these users lives in/usr
. If a subsequent image drops a user record, then it also vanishes from the system. - Using
nss-altfiles
With
nss-altfiles
, you can remove thesystemd
JSON user records. It splits system users into/usr/lib/passwd
and/usr/lib/group
, aligning with the way the OSTree project handles the 3 way merge for/etc
as it relates to/etc/passwd
. Currently, if the/etc/passwd
file is modified in any way on the local system, then subsequent changes to/etc/passwd
in the container image are not applied.Base images built by
rpm-ostree
havenns-altfiles
enabled by default.Also, base images have a system users pre-allocated and managed by the NSS file to avoid UID or GID drift.
In a derived container build, you can also append users to
/usr/lib/passwd
, for example. Usesysusers.d
orDynamicUser=yes
.- Machine-local state for users
The filesystem layout depends on the base image.
By default, the user data is stored in both
/etc
,/etc/passwd
,/etc/shadow
andgroups
, and/home
, depending on the base image. However, the generic base images have to both be machine-local persistent state. In this model/home
is a symlink to/var/home/user
.- Injecting users and SSH keys at system provisioning time
For base images where
/etc
and/var
are configured to persist by default, you can inject users by using installers such as Anaconda or Kickstart.Typically, generic installers are designed for one time bootstrap. Then, the configuration becomes a mutable machine-local state that you can change in Day 2 operations, by using some other mechanism.
You can use the Anaconda installer to set the initial password. However, changing this initial password requires a different in-system tool, such as
passwd
.These flows work equivalently in a
bootc-compatible
system, to support users directly installing generic base images, without requiring changes to the different in-system tool.- Transient home directories
Many operating system deployments minimize persistent, mutable, and executable state. This can damage user home directories.
The
/home
directory can be set astmpfs
, to ensure that user data is cleared across reboots. This approach works especially well when combined with a transient/etc
directory.To set up the user’s home directory to, for example, inject SSH
authorized_keys
or other files, use thesystemd
tmpfiles.d
snippets:f~ /home/user/.ssh/authorized_keys 600 user user - <base64 encoded data>
SSH is embedded in the image as:
/usr/lib/tmpfiles.d/<username-keys.conf
. Another example is a service embedded in the image that can fetch keys from the network and write them. This is the pattern used bycloud-init
.- UID and GID drift
-
The
/etc/passwd
and similar files are a mapping between names and numeric identifiers. When the mapping is dynamic and mixed with "stateless" container image builds, it can cause issues. Each container image build might result in the UID changing due to RPM installation ordering or other reasons. This can be a problem if that user maintains a persistent state. To handle such cases, convert it to usesysusers.d
or useDynamicUser=yes
.
12.2. Injecting secrets in image mode for RHEL
Image mode for RHEL does not have an opinionated mechanism for secrets. You can inject container pull secrets in your system for some cases, for example:
For
bootc
to fetch updates from a registry that requires authentication, you must include a pull secret in a file. In the following example, thecreds
secret contains the registry pull secret.FROM registry.redhat.io/rhel9/bootc-image-builder:latest COPY containers-auth.conf /usr/lib/tmpfiles.d/link-podman-credentials.conf RUN --mount=type=secret,id=creds,required=true cp /run/secrets/creds /usr/lib/container-auth.json && \ chmod 0600 /usr/lib/container-auth.json && \ ln -sr /usr/lib/container-auth.json /etc/ostree/auth.json
To build it, run
podman build --secret id=creds,src=$HOME/.docker/config.json
. Use a single pull secret forbootc
and Podman by using a symlink to both locations to a common persistent file embedded in the container image, for example/usr/lib/container-auth.json
.For Podman to fetch container images, include a pull secret to
/etc/containers/auth.json
. With this configuration, the two stacks share the/usr/lib/container-auth.json
file.- Injecting secrets by embedding them in a container build
- You can include secrets in the container image if the registry server is suitably protected. In some cases, embedding only bootstrap secrets into the container image is a viable pattern, especially alongside a mechanism for having a machine authenticate to a cluster. In this pattern, a provisioning tool, whether run as part of the host system or a container image, uses the bootstrap secret to inject or update other secrets, such as SSH keys, certificates, among others.
- Injecting secrets by using cloud metadata
-
Most production Infrastructure as a Service (IaaS) systems support a metadata server or equivalent which can securely host secrets, particularly bootstrap secrets. Your container image can include tools such as
cloud-init
orignition
to fetch these secrets. - Injecting secrets by embedding them in disk images
-
You can embed
bootstrap secrets
only in disk images. For example, when you generate a cloud disk image from an input container image, such as AMI or OpenStack, the disk image can contain secrets that are effectively machine-local state. Rotating them requires an additional management tool or refreshing the disk images. - Injecting secrets by using bare metal installers
- Installer tools usually support injecting configuration through secrets.
- Injecting secrets through
systemd
credentials -
The
systemd
project has a credential concept for securely acquiring and passing credential data to systems and services, which applies in some deployment methodologies. See the systemd credentials documentation for more details.
Additional resources
12.3. Configuring container pull secrets
To be able to fetch container images, you must configure a host system with a "pull secret", which includes the host updates itself. See the appendix about Injecting secrets in image mode for RHEL documentation for more details.
You can configure the container pull secrets to an image already built. If you use an external installer such as Anaconda for bare metal, or bootc-image-builder
, you must configure the systems with any applicable pull secrets.
The host bootc updates write the configuration to the /etc/ostree/auth.json
file, which is shared with rpm-ostree
.
Podman does not have system wide credentials. Podman accepts the containers-auth
locations that are underneath the following directories:
-
/run
: The content of this directory vanishes on reboot, which is not desired. -
/root
: Part of root’s home directory, which is local mutable state by default.
To unify bootc
and Podman credentials, use a single default global pull secret for both bootc
and Podman. The following container build is an example to unify the bootc
and the Podman credentials. The example expects a secret named creds
to contain the registry pull secret to build.
Prerequisites
- TBD
Procedure
-
Create a symbolic link between
bootc
and Podman to use a single pull secret. By creating the symbolic link, you ensure that both locations are present to a common persistent file embedded in the container image. Create the
/usr/lib/container-auth.json
file.FROM quay.io/<namespace>/<image>:_<tag>_ COPY containers-auth.conf /usr/lib/tmpfiles.d/link-podman-credentials.conf RUN --mount=type=secret,id=creds,required=true cp /run/secrets/creds /usr/lib/container-auth.json && \ chmod 0600 /usr/lib/container-auth.json && \ ln -sr /usr/lib/container-auth.json /etc/ostree/auth.json
When you run the Containerfile, the following actions happen:
-
The Containerfile makes
/run/containers/0/auth.json
a transient runtime file. -
It creates a symbolic link to the
/usr/lib/container-auth.json
. -
It also creates a persistent file, which is also symbolic linked from
/etc/ostree/auth.json
.
-
The Containerfile makes
12.4. Injecting pull secrets for registries and disabling TLS
You can configure container images, pull secrets, and disable TLS for a registry within a system. These actions enable containerized environments to pull images from private or insecure registries.
You can include container pull secrets and other configuration to access a registry inside the base image. However, when installing by using Anaconda, the installation environment might need a duplicate copy of "bootstrap" configuration to access the targeted registry when fetching over the network.
To perform arbitrary changes to the installation environment before the target bootc container image is fetched, you can use the Anaconda %pre
command.
See the containers-auth.json(5)
for more detailed information about format and configurations of the auth.json
file.
Procedure
Configure a pull secret:
%pre mkdir -p /etc/ostree cat > /etc/ostree/auth.json << 'EOF' { "auths": { "quay.io": { "auth": "<your secret here>" } } } EOF %end
With this configuration, the system pulls images from
quay.io
using the provided authentication credentials, which are stored in/etc/ostree/auth.json
.Disable TLS for an insecure registry:
%pre mkdir -p /etc/containers/registries.conf.d/ cat > /etc/containers/registries.conf.d/local-registry.conf << 'EOF' [[registry]] location="[IP_Address]:5000" insecure=true EOF %end
With this configuration, the system pulls container images from a registry that is not secured with TLS. You can use it in development or internal networks.
You can also use %pre
to:
-
Fetch data from the network by using binaries included in the installation environment, such as
curl
. -
Inject trusted certificate authorities into the installation environment
/etc/pki/ca-trust/source/anchors
by using theupdate-ca-trust
command.
You can configure insecure registries similarly by modifying the /etc/containers
directory.
Additional resources
Chapter 13. Appendix: System configuration
13.1. Transient runtime reconfiguration
You can perform a dynamic reconfiguration in the base image configuration. For example, you can run the firewall-cmd --permanent
command to achieve persistent changes across a reboot.
The /etc
directory is persistent by default. If you perform changes made by using tools, for example firewall-cmd --permanent
, the contents of the /etc
on the system can differ from the one described in the container image.
In the default configuration, first make the changes in the base image, then queue the changes without restarting running systems, and then simultaneously write to apply the changes to existing systems only in memory.
You can configure the /etc
directory to be transient by using bind mounts. In this case, the etc
directory is a part of the machine’s local root filesystem. For example, if you inject static IP addresses by using Anaconda kickstarts, they persist across upgrades.
A 3-way merge is applied across upgrades and each "deployment" has its own copy of /etc
.
- The
/run
directory -
The
/run
directory is an API filesystem that is defined to be deleted when the system is restarted. Use the/run
directory for transient files. - Dynamic reconfiguration models
- In the Pull model, you can include code directly embedded in your base image or a privileged container that contacts the remote network server for configuration, and subsequently launch additional container images, by using the Podman API.
In the Push model, some workloads are implemented by tooling such as Ansible.
- systemd
-
You can use systemd units for dynamic transient reconfiguration by writing to
/run/systemd
directory. For example, thesystemctl edit --runtime myservice.service
dynamically changes the configuration of themyservice.service
unit, without persisting the changes. - NetworkManager
-
Use a
/run/NetworkManager/conf.d
directory for applying temporary network configuration. Use thenmcli connection modify --temporary
command to write changes only in memory. Without the--temporary
option, the command writes persistent changes. - podman
-
Use the
podman run --rm
command to automatically remove the container when it exits. Without the--rm
option, thepodman run
command creates a container that persists across system reboots.
13.2. Using dnf
The rhel9/rhel-bootc
container image includes dnf
. There are several use cases:
- Using
dnf
as a part of a container build -
You can use the
RUN dnf install
directive in the Containerfile. - Using
dnf
at runtime
The functionality depends on the dnf
version. You might get an error: error: can’t create transaction lock on /usr/share/rpm/.rpm.lock (Read-only file system)
.
You can use the bootc-usr-overlay
command to create a writable overlay filesystem for /usr
directory. The dnf install
writes to this overlay. You can use this feature for installing debugging tools. Note that changes will be lost on reboot.
- Configuring storage
The supported storage technologies are the following:
-
xfs
/ext4
- Logical volume management (LVM)
- Linux Unified Key Setup (LUKS)
-
You can add other storage packages to the host system.
-
Storage with
bootc-image-builder
You can use the bootc-image-builder tool to create a disk image. The available configuration for partitioning and layout is relatively fixed. The default filesystem type is derived from the container image’s bootc install configuration.
Storage with bootc install You can use the bootc install to-disk
command for flat storage configurations and bootc install to-filesytem
command for more advanced installations. For more information see Advanced installation with to-filesystem
.
13.3. Setting a hostname
To set a custom hostname for your system, modify the /etc/hostname
file. You can set the hostname by using Anaconda, or with a privileged container.
Once you boot a system, you can verify the hostname by using the hostnamectl
command.
13.4. Proxied Internet Access
If you are deploying to an environment requiring internet access by using a proxy, you need to configure services so that they can access resources as intended.
This is done by defining a single file with required environment variables in your configuration, and to reference this by using systemd
drop-in unit files for all such services.
- Defining common proxy environment variables
- This common file has to be subsequently referenced explicitly by each service that requires internet access.
# /etc/example-proxy.env https_proxy="http://example.com:8080" all_proxy="http://example.com:8080" http_proxy="http://example.com:8080" HTTP_PROXY="http://example.com:8080" HTTPS_PROXY="http://example.com:8080" no_proxy="*.example.com,127.0.0.1,0.0.0.0,localhost"
- Defining drop-in units for core services
-
The
bootc
andpodman
tools commonly need proxy configuration. At the current time,bootc
does not always run as asystemd
unit.
# /usr/lib/systemd/system/bootc-fetch-apply-updates.service.d/99-proxy.conf [Service] EnvironmentFile=/etc/example-proxy.env
- Defining proxy use for podman
systemd
units Using the Podman
systemd
configuration, similarly addEnvironmentFile=/etc/example-proxy.env
.You can set the configuration for proxy and environment settings of
podman
and containers in the/etc/containers/containers.conf
configuration file as a root user or in the$HOME/.config/containers/containers.conf
configuration file as a non-root user.
Chapter 14. Appendix: Getting the source code of container images
You can find the source code for bootc image in the Red Hat Ecosystem Catalog.
Procedure
-
Access the Red Hat Ecosystem Catalog and search for
rhel-bootc
. - In the Get this image tab, click Get the source and follow the instructions.
After you extract the content, the input RPM package list and other content resources are available in the
extra_src_dir
directory.The .tar files are snapshots of the input git repository, and contain YAML files with the package lists.
Chapter 15. Appendix: Contributing to the upstream projects
You can contribute to the following upstream bootc projects:
- The upstream git repository is in CentOS Stream.
- The CentOS Stream sources primarily track the Fedora upstream project.