Object Storage

Object storage exposes an S3 API to the storage cluster for applications to put and get data.

Prerequisites

This guide assumes you have created a Rook cluster as explained in the main Kubernetes guide

Create an Object Store

NOTE This example requires you to have at least 3 bluestore OSDs each on a different node. This is because the below erasureCoded chunk settings require at least 3 bluestore OSDs and as failureDomain setting to host (default), each OSD needs to be on a different nodes.

Now we will create the object store, which starts the RGW service in the cluster with the S3 API. Specify your desired settings for the object store in the object.yaml. For more details on the settings see the Object Store CRD.

  1. apiVersion: ceph.rook.io/v1
  2. kind: CephObjectStore
  3. metadata:
  4. name: my-store
  5. namespace: rook-ceph
  6. spec:
  7. metadataPool:
  8. failureDomain: host
  9. replicated:
  10. size: 3
  11. dataPool:
  12. failureDomain: host
  13. erasureCoded:
  14. dataChunks: 2
  15. codingChunks: 1
  16. gateway:
  17. type: s3
  18. sslCertificateRef:
  19. port: 80
  20. securePort:
  21. instances: 1

When the object store is created the Rook operator will create all the pools and other resources necessary to start the service. This may take a minute to complete.

  1. # Create the object store
  2. kubectl create -f object.yaml
  3. # To confirm the object store is configured, wait for the rgw pod to start
  4. kubectl -n rook-ceph get pod -l app=rook-ceph-rgw

Create a User

Next we will create the object store user, which calls the RGW service in the cluster with the S3 API. Specify your desired settings for the object store user in the object-user.yaml. For more details on the settings see the Object Store User CRD.

  1. apiVersion: ceph.rook.io/v1
  2. kind: CephObjectStoreUser
  3. metadata:
  4. name: my-user
  5. namespace: rook-ceph
  6. spec:
  7. store: my-store
  8. displayName: "my display name"

When the object store user is created the Rook operator will create the RGW user on the object store specified, and store the Access Key and Secret Key in a kubernetes secret in the same namespace as the object store user.

  1. # Create the object store user
  2. kubectl create -f object-user.yaml
  3. # To confirm the object store user is configured, describe the secret
  4. kubectl -n rook-ceph describe secret rook-ceph-object-user-my-store-my-user
  5. Name: rook-ceph-object-user-my-store-my-user
  6. Namespace: rook-ceph
  7. Labels: app=rook-ceph-rgw
  8. rook_cluster=rook-ceph
  9. rook_object_store=my-store
  10. Annotations: <none>
  11. Type: kubernetes.io/rook
  12. Data
  13. ====
  14. AccessKey: 20 bytes
  15. SecretKey: 40 bytes

The AccessKey and SecretKey data fields can be mounted in a pod as an environment variable. More information on consuming kubernetes secrets can be found in the K8s secret documentation

To directly retrieve the secrets:

  1. kubectl -n rook-ceph get secret rook-ceph-object-user-my-store-my-user -o yaml | grep AccessKey | awk '{print $2}' | base64 --decode
  2. kubectl -n rook-ceph get secret rook-ceph-object-user-my-store-my-user -o yaml | grep SecretKey | awk '{print $2}' | base64 --decode

Consume the Object Storage

Use an S3 compatible client to create a bucket in the object store.

This section will allow you to test connecting to the object store and uploading and downloading from it. Run the following commands after you have connected to the Rook toolbox.

Install s3cmd

To test the object store we will install the s3cmd tool into the toobox pod.

  1. yum --assumeyes install s3cmd

Connection Environment Variables

To simplify the s3 client commands, you will want to set the four environment variables for use by your client (ie. inside the toolbox):

  1. export AWS_HOST=<host>
  2. export AWS_ENDPOINT=<endpoint>
  3. export AWS_ACCESS_KEY_ID=<accessKey>
  4. export AWS_SECRET_ACCESS_KEY=<secretKey>
  • Host: The DNS host name where the rgw service is found in the cluster. Assuming you are using the default rook-ceph cluster, it will be rook-ceph-rgw-my-store.rook-ceph.
  • Endpoint: The endpoint where the rgw service is listening. Run kubectl -n rook-ceph get svc rook-ceph-rgw-my-store, then combine the clusterIP and the port.
  • Access key: The user’s access_key as printed above
  • Secret key: The user’s secret_key as printed above

The variables for the user generated in this example would be:

  1. export AWS_HOST=rook-ceph-rgw-my-store.rook-ceph
  2. export AWS_ENDPOINT=10.104.35.31:80
  3. export AWS_ACCESS_KEY_ID=XEZDB3UJ6X7HVBE7X7MA
  4. export AWS_SECRET_ACCESS_KEY=7yGIZON7EhFORz0I40BFniML36D2rl8CQQ5kXU6l

The access key and secret key can be retrieved as described in the section above on creating a user.

Create a bucket

Now that the user connection variables were set above, we can proceed to perform operations such as creating buckets.

Create a bucket in the object store

  1. s3cmd mb --no-ssl --host=${AWS_HOST} --host-bucket= s3://rookbucket

List buckets in the object store

  1. s3cmd ls --no-ssl --host=${AWS_HOST}

PUT or GET an object

Upload a file to the newly created bucket

  1. echo "Hello Rook" > /tmp/rookObj
  2. s3cmd put /tmp/rookObj --no-ssl --host=${AWS_HOST} --host-bucket= s3://rookbucket

Download and verify the file from the bucket

  1. s3cmd get s3://rookbucket/rookObj /tmp/rookObj-download --no-ssl --host=${AWS_HOST} --host-bucket=
  2. cat /tmp/rookObj-download

Access External to the Cluster

Rook sets up the object storage so pods will have access internal to the cluster. If your applications are running outside the cluster, you will need to setup an external service through a NodePort.

First, note the service that exposes RGW internal to the cluster. We will leave this service intact and create a new service for external access.

  1. $ kubectl -n rook-ceph get service rook-ceph-rgw-my-store
  2. NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE
  3. rook-ceph-rgw-my-store 10.3.0.177 <none> 80/TCP 2m

Save the external service as rgw-external.yaml:

  1. apiVersion: v1
  2. kind: Service
  3. metadata:
  4. name: rook-ceph-rgw-my-store-external
  5. namespace: rook-ceph
  6. labels:
  7. app: rook-ceph-rgw
  8. rook_cluster: rook-ceph
  9. rook_object_store: my-store
  10. spec:
  11. ports:
  12. - name: rgw
  13. port: 80
  14. protocol: TCP
  15. targetPort: 80
  16. selector:
  17. app: rook-ceph-rgw
  18. rook_cluster: rook-ceph
  19. rook_object_store: my-store
  20. sessionAffinity: None
  21. type: NodePort

Now create the external service.

  1. kubectl create -f rgw-external.yaml

See both rgw services running and notice what port the external service is running on:

  1. $ kubectl -n rook-ceph get service rook-ceph-rgw-my-store rook-ceph-rgw-my-store-external
  2. NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
  3. rook-ceph-rgw-my-store ClusterIP 10.104.82.228 <none> 80/TCP 4m
  4. rook-ceph-rgw-my-store-external NodePort 10.111.113.237 <none> 80:31536/TCP 39s

Internally the rgw service is running on port 80. The external port in this case is 31536. Now you can access the object store from anywhere! All you need is the hostname for any machine in the cluster, the external port, and the user credentials.