You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: hack/inventory/production-sample/rook/cluster.yaml
+49-25Lines changed: 49 additions & 25 deletions
Original file line number
Diff line number
Diff line change
@@ -16,13 +16,13 @@ metadata:
16
16
spec:
17
17
cephVersion:
18
18
# The container image used to launch the Ceph daemon pods (mon, mgr, osd, mds, rgw).
19
-
#v14 is nautilus, v15 is octopus, and v16 is pacific.
19
+
# v15 is octopus, and v16 is pacific.
20
20
# RECOMMENDATION: In production, use a specific version tag instead of the general v14 flag, which pulls the latest release and could result in different
21
21
# versions running within the cluster. See tags available at https://hub.docker.com/r/ceph/ceph/tags/.
22
-
# If you want to be more precise, you can always use a timestamp tag such quay.io/ceph/ceph:v16.2.6-20210918
22
+
# If you want to be more precise, you can always use a timestamp tag such quay.io/ceph/ceph:v16.2.10-20220721
23
23
# This tag might not contain a new Ceph version, just security fixes from the underlying operating system, which will reduce vulnerabilities
24
-
image: quay.io/ceph/ceph:v15.2.8
25
-
# Whether to allow unsupported versions of Ceph. Currently `nautilus`, `octopus`, and `pacific` are supported.
24
+
image: quay.io/ceph/ceph:v16.2.10
25
+
# Whether to allow unsupported versions of Ceph. Currently `octopus` and `pacific` are supported.
26
26
# Future versions such as `pacific` would require this to be set to `true`.
27
27
# Do not set to true in production.
28
28
allowUnsupported: false
@@ -33,13 +33,13 @@ spec:
33
33
# Whether or not upgrade should continue even if a check fails
34
34
# This means Ceph's status could be degraded and we don't recommend upgrading but you might decide otherwise
35
35
# Use at your OWN risk
36
-
# To understand Rook's upgrade process of Ceph, read https://rook.io/docs/rook/master/ceph-upgrade.html#ceph-version-upgrades
36
+
# To understand Rook's upgrade process of Ceph, read https://rook.io/docs/rook/latest/ceph-upgrade.html#ceph-version-upgrades
37
37
skipUpgradeChecks: false
38
38
# Whether or not continue if PGs are not clean during an upgrade
39
39
continueUpgradeAfterChecksEvenIfNotHealthy: false
40
40
# WaitTimeoutForHealthyOSDInMinutes defines the time (in minutes) the operator would wait before an OSD can be stopped for upgrade or restart.
41
41
# If the timeout exceeds and OSD is not ok to stop, then the operator would skip upgrade for the current OSD and proceed with the next one
42
-
# if `continueUpgradeAfterChecksEvenIfNotHealthy` is `false`. If `continueUpgradeAfterChecksEvenIfNotHealthy` is `true`, then opertor would
42
+
# if `continueUpgradeAfterChecksEvenIfNotHealthy` is `false`. If `continueUpgradeAfterChecksEvenIfNotHealthy` is `true`, then operator would
43
43
# continue with the upgrade of an OSD even if its not ok to stop after the timeout. This timeout won't be applied if `skipUpgradeChecks` is `true`.
44
44
# The default wait timeout is 10 minutes.
45
45
waitTimeoutForHealthyOSDInMinutes: 10
@@ -54,7 +54,8 @@ spec:
54
54
# When higher availability of the mgr is needed, increase the count to 2.
55
55
# In that case, one mgr will be active and one in standby. When Ceph updates which
56
56
# mgr is active, Rook will update the mgr services to match the active mgr.
57
-
count: 1
57
+
count: 2
58
+
allowMultiplePerNode: false
58
59
modules:
59
60
# Several modules should not need to be included in this list. The "dashboard" and "monitoring" modules
60
61
# are already enabled by other settings in the cluster CR.
@@ -73,13 +74,20 @@ spec:
73
74
monitoring:
74
75
# requires Prometheus to be pre-installed
75
76
enabled: false
76
-
# namespace to deploy prometheusRule in. If empty, namespace of the cluster will be used.
77
-
# Recommended:
78
-
# If you have a single rook-ceph cluster, set the rulesNamespace to the same namespace as the cluster or keep it empty.
79
-
# If you have multiple rook-ceph clusters in the same k8s cluster, choose the same namespace (ideally, namespace with prometheus
80
-
# deployed) to set rulesNamespace for all the clusters. Otherwise, you will get duplicate alerts with multiple alert definitions.
81
-
rulesNamespace: rook-ceph
82
77
network:
78
+
connections:
79
+
# Whether to encrypt the data in transit across the wire to prevent eavesdropping the data on the network.
80
+
# The default is false. When encryption is enabled, all communication between clients and Ceph daemons, or between Ceph daemons will be encrypted.
81
+
# When encryption is not enabled, clients still establish a strong initial authentication and data integrity is still validated with a crc check.
82
+
# IMPORTANT: Encryption requires the 5.11 kernel for the latest nbd and cephfs drivers. Alternatively for testing only,
83
+
# you can set the "mounter: rbd-nbd" in the rbd storage class, or "mounter: fuse" in the cephfs storage class.
84
+
# The nbd and fuse drivers are *not* recommended in production since restarting the csi driver pod will disconnect the volumes.
85
+
encryption:
86
+
enabled: false
87
+
# Whether to compress the data in transit across the wire. The default is false.
88
+
# Requires Ceph Quincy (v17) or newer. Also see the kernel requirements above for encryption.
89
+
compression:
90
+
enabled: false
83
91
# enable host networking
84
92
#provider: host
85
93
# enable the Multus network provider
@@ -107,8 +115,9 @@ spec:
107
115
# enable log collector, daemons will log on files and rotate
108
116
# logCollector:
109
117
# enabled: true
110
-
# periodicity: 24h # SUFFIX may be 'h' for hours or 'd' for days.
111
-
# automate [data cleanup process](https://github.com/rook/rook/blob/master/Documentation/ceph-teardown.md#delete-the-data-on-hosts) in cluster destruction.
118
+
# periodicity: daily # one of: hourly, daily, weekly, monthly
119
+
# maxLogSize: 500M # SUFFIX may be 'M' or 'G'. Must be at least 1M.
120
+
# automate [data cleanup process](https://github.com/rook/rook/blob/master/Documentation/Storage-Configuration/ceph-teardown.md#delete-the-data-on-hosts) in cluster destruction.
112
121
cleanupPolicy:
113
122
# Since cluster cleanup is destructive to data, confirmation is required.
114
123
# To destroy all Rook data on hosts during uninstall, confirmation must be set to "yes-really-destroy-data".
@@ -158,6 +167,7 @@ spec:
158
167
# or when AllowMultiplePerNode is false. Otherwise this anti-affinity rule is a
159
168
# preferred rule with weight: 50.
160
169
# osd:
170
+
# prepareosd:
161
171
# mgr:
162
172
# cleanup:
163
173
annotations:
@@ -166,6 +176,10 @@ spec:
166
176
# osd:
167
177
# cleanup:
168
178
# prepareosd:
179
+
# clusterMetadata annotations will be applied to only `rook-ceph-mon-endpoints` configmap and the `rook-ceph-mon` and `rook-ceph-admin-keyring` secrets.
180
+
# And clusterMetadata annotations will not be merged with `all` annotations.
181
+
# clusterMetadata:
182
+
# kubed.appscode.com/sync: "true"
169
183
# If no mgr annotations are set, prometheus scrape annotations will be set by default.
170
184
# mgr:
171
185
labels:
@@ -178,22 +192,23 @@ spec:
178
192
# monitoring is a list of key-value pairs. It is injected into all the monitoring resources created by operator.
179
193
# These labels can be passed as LabelSelector to Prometheus
180
194
# monitoring:
195
+
# crashcollector:
181
196
resources:
182
197
# The requests and limits set here, allow the mgr pod to use half of one CPU core and 1 gigabyte of memory
183
198
# mgr:
184
199
# limits:
185
-
# cpu: "1"
200
+
# cpu: "500m"
186
201
# memory: "1024Mi"
187
202
# requests:
188
-
# cpu: "1"
203
+
# cpu: "500m"
189
204
# memory: "1024Mi"
190
205
# The above example requests/limits can also be added to the other components
191
206
# mon:
192
207
# limits:
193
-
# cpu: "2"
208
+
# cpu: "1"
194
209
# memory: "2048Mi"
195
210
# requests:
196
-
# cpu: "2"
211
+
# cpu: "1"
197
212
# memory: "2048Mi"
198
213
# osd:
199
214
# limits:
@@ -214,10 +229,11 @@ spec:
214
229
# The option to automatically remove OSDs that are out and are safe to destroy.
0 commit comments