You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: hack/inventory/production-sample/rook/cluster.yaml
+40-27Lines changed: 40 additions & 27 deletions
Original file line number
Diff line number
Diff line change
@@ -16,14 +16,14 @@ metadata:
16
16
spec:
17
17
cephVersion:
18
18
# The container image used to launch the Ceph daemon pods (mon, mgr, osd, mds, rgw).
19
-
#v15 is octopus, and v16 is pacific.
20
-
# RECOMMENDATION: In production, use a specific version tag instead of the general v14 flag, which pulls the latest release and could result in different
19
+
#v16 is Pacific, and v17 is Quincy.
20
+
# RECOMMENDATION: In production, use a specific version tag instead of the general v17 flag, which pulls the latest release and could result in different
21
21
# versions running within the cluster. See tags available at https://hub.docker.com/r/ceph/ceph/tags/.
22
-
# If you want to be more precise, you can always use a timestamp tag such quay.io/ceph/ceph:v16.2.10-20220721
22
+
# If you want to be more precise, you can always use a timestamp tag such quay.io/ceph/ceph:v17.2.6-20230410
23
23
# This tag might not contain a new Ceph version, just security fixes from the underlying operating system, which will reduce vulnerabilities
24
-
image: quay.io/ceph/ceph:v16.2.10
25
-
# Whether to allow unsupported versions of Ceph. Currently `octopus` and `pacific` are supported.
26
-
# Future versions such as `pacific` would require this to be set to `true`.
24
+
image: quay.io/ceph/ceph:v17.2.6
25
+
# Whether to allow unsupported versions of Ceph. Currently `pacific`, `quincy`, and `reef` are supported.
26
+
# Future versions such as `squid` (v19) would require this to be set to `true`.
27
27
# Do not set to true in production.
28
28
allowUnsupported: false
29
29
# The path on the host where configuration files will be persisted. Must be specified.
# Whether SSL should be verified if the Prometheus server is using https
76
+
# prometheusEndpointSSLVerify: false
73
77
# enable prometheus alerting for cluster
74
78
monitoring:
75
79
# requires Prometheus to be pre-installed
76
80
enabled: false
81
+
# Whether to disable the metrics reported by Ceph. If false, the prometheus mgr module and Ceph exporter are enabled.
82
+
# If true, the prometheus mgr module and Ceph exporter are both disabled. Default is false.
83
+
metricsDisabled: false
77
84
network:
78
85
connections:
79
86
# Whether to encrypt the data in transit across the wire to prevent eavesdropping the data on the network.
@@ -88,35 +95,45 @@ spec:
88
95
# Requires Ceph Quincy (v17) or newer. Also see the kernel requirements above for encryption.
89
96
compression:
90
97
enabled: false
98
+
# Whether to require communication over msgr2. If true, the msgr v1 port (6789) will be disabled
99
+
# and clients will be required to connect to the Ceph cluster with the v2 port (3300).
100
+
# Requires a kernel that supports msgr v2 (kernel 5.11 or CentOS 8.4 or newer).
101
+
requireMsgr2: false
91
102
# enable host networking
92
103
#provider: host
93
104
# enable the Multus network provider
94
105
#provider: multus
95
106
#selectors:
96
-
# The selector keys are required to be `public` and `cluster`.
97
-
# Based on the configuration, the operator will do the following:
98
-
# 1. if only the `public` selector key is specified both public_network and cluster_network Ceph settings will listen on that interface
99
-
# 2. if both `public` and `cluster` selector keys are specified the first one will point to 'public_network' flag and the second one to 'cluster_network'
100
-
#
101
-
# In order to work, each selector value must match a NetworkAttachmentDefinition object in Multus
102
-
#
103
-
#public: public-conf --> NetworkAttachmentDefinition object name in Multus
104
-
#cluster: cluster-conf --> NetworkAttachmentDefinition object name in Multus
107
+
# The selector keys are required to be `public` and `cluster`.
108
+
# Based on the configuration, the operator will do the following:
109
+
# 1. if only the `public` selector key is specified both public_network and cluster_network Ceph settings will listen on that interface
110
+
# 2. if both `public` and `cluster` selector keys are specified the first one will point to 'public_network' flag and the second one to 'cluster_network'
111
+
#
112
+
# In order to work, each selector value must match a NetworkAttachmentDefinition object in Multus
113
+
#
114
+
#public: public-conf --> NetworkAttachmentDefinition object name in Multus
115
+
#cluster: cluster-conf --> NetworkAttachmentDefinition object name in Multus
105
116
# Provide internet protocol version. IPv6, IPv4 or empty string are valid options. Empty string would mean IPv4
106
117
#ipFamily: "IPv6"
107
118
# Ceph daemons to listen on both IPv4 and Ipv6 networks
108
119
#dualStack: false
120
+
# Enable multiClusterService to export the mon and OSD services to peer cluster.
121
+
# This is useful to support RBD mirroring between two clusters having overlapping CIDRs.
122
+
# Ensure that peer clusters are connected using an MCS API compatible application, like Globalnet Submariner.
123
+
#multiClusterService:
124
+
# enabled: false
125
+
109
126
# enable the crash collector for ceph daemon crash collection
110
127
crashCollector:
111
128
disable: false
112
129
# Uncomment daysToRetain to prune ceph crash entries older than the
113
130
# specified number of days.
114
131
#daysToRetain: 30
115
132
# enable log collector, daemons will log on files and rotate
116
-
#logCollector:
117
-
#enabled: true
118
-
# periodicity: daily # one of: hourly, daily, weekly, monthly
119
-
#maxLogSize: 500M # SUFFIX may be 'M' or 'G'. Must be at least 1M.
133
+
logCollector:
134
+
enabled: false
135
+
periodicity: daily # one of: hourly, daily, weekly, monthly
136
+
maxLogSize: 500M# SUFFIX may be 'M' or 'G'. Must be at least 1M.
120
137
# automate [data cleanup process](https://github.com/rook/rook/blob/master/Documentation/Storage-Configuration/ceph-teardown.md#delete-the-data-on-hosts) in cluster destruction.
121
138
cleanupPolicy:
122
139
# Since cluster cleanup is destructive to data, confirmation is required.
@@ -242,11 +259,10 @@ spec:
242
259
# crushRoot: "custom-root" # specify a non-default root label for the CRUSH map
243
260
# metadataDevice: "md0" # specify a non-rotational storage so ceph-volume will use it as block db device of bluestore.
244
261
# databaseSizeMB: "1024" # uncomment if the disks are smaller than 100 GB
245
-
# journalSizeMB: "1024" # uncomment if the disks are 20 GB or smaller
246
262
# osdsPerDevice: "1" # this value can be overridden at the node or device level
247
263
# encryptedDevice: "true" # the default value for this option is "false"
248
-
# Individual nodes and their config can be specified as well, but 'useAllNodes' above must be set to false. Then, only the named
249
-
# nodes below will be used as storage resources. Each node's 'name' field should match their 'kubernetes.io/hostname' label.
264
+
# Individual nodes and their config can be specified as well, but 'useAllNodes' above must be set to false. Then, only the named
265
+
# nodes below will be used as storage resources. Each node's 'name' field should match their 'kubernetes.io/hostname' label.
250
266
# nodes:
251
267
# - name: "172.17.4.201"
252
268
# devices: # specific devices to use for storage can be specified for each node
@@ -260,6 +276,8 @@ spec:
260
276
# deviceFilter: "^sd."
261
277
# when onlyApplyOSDPlacement is false, will merge both placement.All() and placement.osd
262
278
onlyApplyOSDPlacement: false
279
+
# Time for which an OSD pod will sleep before restarting, if it stopped due to flapping
280
+
# flappingRestartIntervalHours: 24
263
281
# The section for configuring management of daemon disruptions during upgrade or fencing.
264
282
disruptionManagement:
265
283
# If true, the operator will create and manage PodDisruptionBudgets for OSD, Mon, RGW, and MDS daemons. OSD PDBs are managed dynamically
@@ -273,11 +291,6 @@ spec:
273
291
# Operator will continue with the next drain if the timeout exceeds. It only works if `managePodBudgets` is `true`.
274
292
# No values or 0 means that the operator will wait until the placement groups are healthy before unblocking the next drain.
275
293
pgHealthCheckTimeout: 0
276
-
# If true, the operator will create and manage MachineDisruptionBudgets to ensure OSDs are only fenced when the cluster is healthy.
277
-
# Only available on OpenShift.
278
-
manageMachineDisruptionBudgets: false
279
-
# Namespace in which to watch for the MachineDisruptionBudgets.
0 commit comments