From 2ffc44f6b45fdc88affec57cfebd3f089570f740 Mon Sep 17 00:00:00 2001 From: Travis Nielsen Date: Thu, 2 Jan 2025 08:45:44 -0700 Subject: [PATCH] build: set release version to v1.16.1 For the patch release update the docs and examples to v1.16.1. Signed-off-by: Travis Nielsen --- Documentation/Getting-Started/quickstart.md | 2 +- .../Monitoring/ceph-monitoring.md | 2 +- Documentation/Upgrade/rook-upgrade.md | 32 +++++++++---------- deploy/charts/rook-ceph/values.yaml | 2 +- deploy/examples/direct-mount.yaml | 2 +- deploy/examples/images.txt | 2 +- deploy/examples/operator-openshift.yaml | 2 +- deploy/examples/operator.yaml | 2 +- deploy/examples/osd-purge.yaml | 2 +- deploy/examples/toolbox-job.yaml | 4 +-- deploy/examples/toolbox-operator-image.yaml | 2 +- 11 files changed, 27 insertions(+), 27 deletions(-) diff --git a/Documentation/Getting-Started/quickstart.md b/Documentation/Getting-Started/quickstart.md index 13597ca26320..a9b7d397abf9 100644 --- a/Documentation/Getting-Started/quickstart.md +++ b/Documentation/Getting-Started/quickstart.md @@ -36,7 +36,7 @@ To configure the Ceph storage cluster, at least one of these local storage optio A simple Rook cluster is created for Kubernetes with the following `kubectl` commands and [example manifests](https://github.com/rook/rook/blob/master/deploy/examples). ```console -$ git clone --single-branch --branch v1.16.0 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.16.1 https://github.com/rook/rook.git cd rook/deploy/examples kubectl create -f crds.yaml -f common.yaml -f operator.yaml kubectl create -f cluster.yaml diff --git a/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md b/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md index 34f0582bcc3c..79bf9b0c66c2 100644 --- a/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md +++ b/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md @@ -48,7 +48,7 @@ There are two sources for metrics collection: From the root of your locally cloned Rook repo, go the monitoring directory: ```console -$ git clone --single-branch --branch v1.16.0 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.16.1 https://github.com/rook/rook.git cd rook/deploy/examples/monitoring ``` diff --git a/Documentation/Upgrade/rook-upgrade.md b/Documentation/Upgrade/rook-upgrade.md index 33f8adf3e6bc..a9b696fca475 100644 --- a/Documentation/Upgrade/rook-upgrade.md +++ b/Documentation/Upgrade/rook-upgrade.md @@ -127,8 +127,8 @@ In order to successfully upgrade a Rook cluster, the following prerequisites mus ## Rook Operator Upgrade -The examples given in this guide upgrade a live Rook cluster running `v1.15.6` to -the version `v1.16.0`. This upgrade should work from any official patch release of Rook v1.15 to any +The examples given in this guide upgrade a live Rook cluster running `v1.15.7` to +the version `v1.16.1`. This upgrade should work from any official patch release of Rook v1.15 to any official patch release of v1.16. Let's get started! @@ -155,7 +155,7 @@ by the Operator. Also update the Custom Resource Definitions (CRDs). Get the latest common resources manifests that contain the latest changes. ```console -git clone --single-branch --depth=1 --branch v1.16.0 https://github.com/rook/rook.git +git clone --single-branch --depth=1 --branch v1.16.1 https://github.com/rook/rook.git cd rook/deploy/examples ``` @@ -194,7 +194,7 @@ The largest portion of the upgrade is triggered when the operator's image is upd When the operator is updated, it will proceed to update all of the Ceph daemons. ```console -kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.16.0 +kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.16.1 ``` ### **3. Update Ceph CSI** @@ -221,18 +221,18 @@ watch --exec kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -l rook_cluster= ``` As an example, this cluster is midway through updating the OSDs. When all deployments report `1/1/1` -availability and `rook-version=v1.16.0`, the Ceph cluster's core components are fully updated. +availability and `rook-version=v1.16.1`, the Ceph cluster's core components are fully updated. ```console Every 2.0s: kubectl -n rook-ceph get deployment -o j... -rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.16.0 -rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.16.0 -rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.16.0 -rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.16.0 -rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.16.0 -rook-ceph-osd-1 req/upd/avl: 1/1/1 rook-version=v1.15.6 -rook-ceph-osd-2 req/upd/avl: 1/1/1 rook-version=v1.15.6 +rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.16.1 +rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.16.1 +rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.16.1 +rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.16.1 +rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.16.1 +rook-ceph-osd-1 req/upd/avl: 1/1/1 rook-version=v1.15.7 +rook-ceph-osd-2 req/upd/avl: 1/1/1 rook-version=v1.15.7 ``` An easy check to see if the upgrade is totally finished is to check that there is only one @@ -241,14 +241,14 @@ An easy check to see if the upgrade is totally finished is to check that there i ```console # kubectl -n $ROOK_CLUSTER_NAMESPACE get deployment -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{"rook-version="}{.metadata.labels.rook-version}{"\n"}{end}' | sort | uniq This cluster is not yet finished: - rook-version=v1.15.6 - rook-version=v1.16.0 + rook-version=v1.15.7 + rook-version=v1.16.1 This cluster is finished: - rook-version=v1.16.0 + rook-version=v1.16.1 ``` ### **5. Verify the updated cluster** -At this point, the Rook operator should be running version `rook/ceph:v1.16.0`. +At this point, the Rook operator should be running version `rook/ceph:v1.16.1`. Verify the CephCluster health using the [health verification doc](health-verification.md). diff --git a/deploy/charts/rook-ceph/values.yaml b/deploy/charts/rook-ceph/values.yaml index 66a954b75c02..da4badd30062 100644 --- a/deploy/charts/rook-ceph/values.yaml +++ b/deploy/charts/rook-ceph/values.yaml @@ -7,7 +7,7 @@ image: repository: docker.io/rook/ceph # -- Image tag # @default -- `master` - tag: v1.16.0 + tag: v1.16.1 # -- Image pull policy pullPolicy: IfNotPresent diff --git a/deploy/examples/direct-mount.yaml b/deploy/examples/direct-mount.yaml index 8b30a0e20f27..bfbeb4cdb207 100644 --- a/deploy/examples/direct-mount.yaml +++ b/deploy/examples/direct-mount.yaml @@ -19,7 +19,7 @@ spec: serviceAccountName: rook-ceph-default containers: - name: rook-direct-mount - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 command: ["/bin/bash"] args: ["-m", "-c", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent diff --git a/deploy/examples/images.txt b/deploy/examples/images.txt index 02a78bb5fd13..4aeff0331ca5 100644 --- a/deploy/examples/images.txt +++ b/deploy/examples/images.txt @@ -1,4 +1,4 @@ - docker.io/rook/ceph:v1.16.0 + docker.io/rook/ceph:v1.16.1 gcr.io/k8s-staging-sig-storage/objectstorage-sidecar:v20240513-v0.1.0-35-gefb3255 quay.io/ceph/ceph:v19.2.0 quay.io/ceph/cosi:v0.1.2 diff --git a/deploy/examples/operator-openshift.yaml b/deploy/examples/operator-openshift.yaml index 0c6c739ed7d1..757a3a3e8b29 100644 --- a/deploy/examples/operator-openshift.yaml +++ b/deploy/examples/operator-openshift.yaml @@ -671,7 +671,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 args: ["ceph", "operator"] securityContext: runAsNonRoot: true diff --git a/deploy/examples/operator.yaml b/deploy/examples/operator.yaml index 270517f3323a..17db01f52de4 100644 --- a/deploy/examples/operator.yaml +++ b/deploy/examples/operator.yaml @@ -600,7 +600,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 args: ["ceph", "operator"] securityContext: runAsNonRoot: true diff --git a/deploy/examples/osd-purge.yaml b/deploy/examples/osd-purge.yaml index 252510e52b29..f366808a2ca8 100644 --- a/deploy/examples/osd-purge.yaml +++ b/deploy/examples/osd-purge.yaml @@ -28,7 +28,7 @@ spec: serviceAccountName: rook-ceph-purge-osd containers: - name: osd-removal - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 # TODO: Insert the OSD ID in the last parameter that is to be removed # The OSD IDs are a comma-separated list. For example: "0" or "0,2". # If you want to preserve the OSD PVCs, set `--preserve-pvc true`. diff --git a/deploy/examples/toolbox-job.yaml b/deploy/examples/toolbox-job.yaml index d147280870e6..c3dd7be8be2e 100644 --- a/deploy/examples/toolbox-job.yaml +++ b/deploy/examples/toolbox-job.yaml @@ -10,7 +10,7 @@ spec: spec: initContainers: - name: config-init - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 command: ["/usr/local/bin/toolbox.sh"] args: ["--skip-watch"] imagePullPolicy: IfNotPresent @@ -29,7 +29,7 @@ spec: mountPath: /var/lib/rook-ceph-mon containers: - name: script - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 volumeMounts: - mountPath: /etc/ceph name: ceph-config diff --git a/deploy/examples/toolbox-operator-image.yaml b/deploy/examples/toolbox-operator-image.yaml index a501e7422ba3..f42cda10f110 100644 --- a/deploy/examples/toolbox-operator-image.yaml +++ b/deploy/examples/toolbox-operator-image.yaml @@ -25,7 +25,7 @@ spec: serviceAccountName: rook-ceph-default containers: - name: rook-ceph-tools-operator-image - image: docker.io/rook/ceph:v1.16.0 + image: docker.io/rook/ceph:v1.16.1 command: - /bin/bash - -c