Ce contenu n'est pas disponible dans la langue sélectionnée.
Chapter 14. Using Persistent Storage in Fuse on OpenShift
Fuse on OpenShift applications are based on OpenShift containers, which do not have a persistent filesystem. Every time you start an application, it is started in a new container with an immutable Docker-formatted image. Hence any persisted data in the file systems is lost when the container stops. But applications need to store some state as data in a persistent store and sometimes applications share access to a common data store. OpenShift platform supports provisioning of external stores as Persistent Storage.
14.1. About volumes and volume types
OpenShift allows pods and containers to mount Volumes as file systems which are backed by multiple local or network attached storage endpoints.
Volume types include:
- emptydir (empty directory): This is a default volume type. It is a directory which gets allocated when the pod is created on a local host. It is not copied across the servers and when you delete the pod the directory is removed.
- configmap: It is a directory with contents populated with key-value pairs from a named configmap.
- hostPath (host directory): It is a directory with specific path on any host and it requires elevated privileges.
- secret (mounted secret): Secret volumes mount a named secret to the provided directory.
- persistentvolumeclaim or pvc (persistent volume claim): This links the volume directory in the container to a persistent volume claim you have allocated by name. A persistent volume claim is a request to allocate storage. Note that if your claim is not bound, your pods will not start.
Volumes are configured at the Pod level and can only directly access an external storage using hostPath. Hence it is harder to mange the access to shared resources for multiple Pods as hostPath volumes.
14.2. About PersistentVolumes
PersistentVolumes allow cluster administrators to provision cluster wide storage which is backed by various types of network storage like NFS, Ceph RBD, AWS Elastic Block Store (EBS), etc. PersistentVolumes also specify capacity, access modes, and recycling policies. This allows pods from multiple Projects to access persistent storage without worrying about the nature of the underlying resource.
See Configuring Persistent Storage for creating various types of PersistentVolumes.
14.3. Configuring persistent volume
You can provision a persistent volume by creating a configuration file. This storage then can be accessed by creating a PersistentVolume Claim.
Procedure
Create a configuration file named
pv.yaml
using the sample configuration below. This provisions a path on the host machine as a PersistentVolume named pv001.apiVersion: v1 kind: PersistentVolume metadata: name: pv0001 spec: accessModes: - ReadWriteOnce capacity: storage: 2Mi hostPath: path: /data/pv0001/
Here the host path is
/data/pv0001
and storage capacity is limited to 2MB. For example, when using OpenShift CDK it will provision the directory/data/pv0001
from the virtual machine hosting the OpenShift Cluster.Create the
PersistentVolume
.oc create -f pv.yaml
Verify the creation of PersistentVolume. This will list all the PersistentVolumes configured in your OpenShift cluster:
oc get pv
14.4. Creating PersistentVolumeClaims
A PersistentVolume exposes a storage endpoint as a named entity in an OpenShift cluster. To access this storage from Projects, PersistentVolumeClaims must be created that can access the PersistentVolume. PersistentVolumeClaims are created for each Project with customized claims for a certain amount of storage with certain access modes.
Procedure
The sample configuration below creates a claim named pvc0001 for 1MB of storage with read-write-once access against a PersistentVolume named pv0001.
apiVersion: v1 kind: PersistentVolumeClaim metadata: name: pvc0001 spec: accessModes: - ReadWriteOnce resources: requests: storage: 1Mi
14.5. Using persistent volumes in pods
Pods use volume mounts to define the filesystem mount location and volumes to define reference PersistentVolumeClaims
.
Procedure
Create a sample container configuration as shown below which mounts
PersistentVolumeClaim
pvc0001 at/usr/share/data
in its filesystem.spec: template: spec: containers: - volumeMounts: - name: vol0001 mountPath: /usr/share/data volumes: - name: vol0001 persistentVolumeClaim: claimName: pvc0001
Any data written by the application to the directory
/usr/share/data
is now persisted across container restarts.Add this configuration in the file
src/main/jkube/deployment.yml
in a Fuse on OpenShift application and create OpenShift resources using command:mvn oc:resource-apply
Verify that the created DeploymentConfiguration has the volume mount and the volume.
oc describe deploymentconfig <application-dc-name>
For Fuse on OpenShift quickstarts, replace the
<application-dc-name>
with the Maven project name, for examplespring-boot-camel
.