mirror of
https://github.com/ceph/ceph-csi.git
synced 2025-06-13 02:33:34 +00:00
cleanup: Move common files to deploy folder
Few common files related to deployments were kept in the examples folder initially. Moving them to deploy folder and updating the relevant files. Signed-off-by: karthik-us <ksubrahm@redhat.com>
This commit is contained in:
21
deploy/ceph-conf.yaml
Normal file
21
deploy/ceph-conf.yaml
Normal file
@ -0,0 +1,21 @@
|
||||
---
|
||||
# This is a sample configmap that helps define a Ceph configuration as required
|
||||
# by the CSI plugins.
|
||||
|
||||
# Sample ceph.conf available at
|
||||
# https://github.com/ceph/ceph/blob/master/src/sample.ceph.conf Detailed
|
||||
# documentation is available at
|
||||
# https://docs.ceph.com/en/latest/rados/configuration/ceph-conf/
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
data:
|
||||
ceph.conf: |
|
||||
[global]
|
||||
auth_cluster_required = cephx
|
||||
auth_service_required = cephx
|
||||
auth_client_required = cephx
|
||||
|
||||
# keyring is a required key and its value should be empty
|
||||
keyring: |
|
||||
metadata:
|
||||
name: ceph-config
|
89
deploy/csi-config-map-sample.yaml
Normal file
89
deploy/csi-config-map-sample.yaml
Normal file
@ -0,0 +1,89 @@
|
||||
---
|
||||
# This is a sample configmap that helps define a Ceph cluster configuration
|
||||
# as required by the CSI plugins.
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
# Lets see the different configuration under config.json key.
|
||||
# The <cluster-id> is used by the CSI plugin to uniquely identify and use a
|
||||
# Ceph cluster, the value MUST match the value provided as `clusterID` in the
|
||||
# StorageClass
|
||||
# The <MONValue#> fields are the various monitor addresses for the Ceph cluster
|
||||
# identified by the <cluster-id>
|
||||
# If a CSI plugin is using more than one Ceph cluster, repeat the section for
|
||||
# each such cluster in use.
|
||||
# To add more clusters or edit MON addresses in an existing configmap, use
|
||||
# the `kubectl replace` command.
|
||||
# The "rbd.rados-namespace" is optional and represents a radosNamespace in the
|
||||
# pool. If any given, all of the rbd images, snapshots, and other metadata will
|
||||
# be stored within the radosNamespace.
|
||||
# NOTE: The given radosNamespace must already exists in the pool.
|
||||
# NOTE: Make sure you don't add radosNamespace option to a currently in use
|
||||
# configuration as it will cause issues.
|
||||
# The field "cephFS.subvolumeGroup" is optional and defaults to "csi".
|
||||
# The "cephFS.netNamespaceFilePath" fields are the various network namespace
|
||||
# path for the Ceph cluster identified by the <cluster-id>, This will be used
|
||||
# by the CephFS CSI plugin to execute the mount -t in the
|
||||
# network namespace specified by the "cephFS.netNamespaceFilePath".
|
||||
# The "nfs.netNamespaceFilePath" fields are the various network namespace
|
||||
# path for the Ceph cluster identified by the <cluster-id>, This will be used
|
||||
# by the NFS CSI plugin to execute the mount -t in the
|
||||
# network namespace specified by the "nfs.netNamespaceFilePath".
|
||||
# The "rbd.netNamespaceFilePath" fields are the various network namespace
|
||||
# path for the Ceph cluster identified by the <cluster-id>, This will be used
|
||||
# by the RBD CSI plugin to execute the rbd map/unmap in the
|
||||
# network namespace specified by the "rbd.netNamespaceFilePath".
|
||||
# If a CSI plugin is using more than one Ceph cluster, repeat the section for
|
||||
# each such cluster in use.
|
||||
# NOTE: Changes to the configmap is automatically updated in the running pods,
|
||||
# thus restarting existing pods using the configmap is NOT required on edits
|
||||
# to the configmap.
|
||||
|
||||
# Lets see the different configuration under cluster-mapping.json key.
|
||||
# This configuration is needed when volumes are mirrored using the Ceph-CSI.
|
||||
# clusterIDMapping holds the mapping between two clusterId's of storage
|
||||
# clusters.
|
||||
# RBDPoolIDMapping holds the mapping between two poolId's of storage clusters.
|
||||
# CephFSFscIDMapping holds the mapping between two FscId's of storage
|
||||
# clusters.
|
||||
data:
|
||||
config.json: |-
|
||||
[
|
||||
{
|
||||
"clusterID": "<cluster-id>",
|
||||
"rbd": {
|
||||
"netNamespaceFilePath": "<kubeletRootPath>/plugins/rbd.csi.ceph.com/net",
|
||||
"radosNamespace": "<rados-namespace>",
|
||||
},
|
||||
"monitors": [
|
||||
"<MONValue1>",
|
||||
"<MONValue2>",
|
||||
...
|
||||
"<MONValueN>"
|
||||
],
|
||||
"cephFS": {
|
||||
"subvolumeGroup": "<subvolumegroup for cephFS volumes>"
|
||||
"netNamespaceFilePath": "<kubeletRootPath>/plugins/cephfs.csi.ceph.com/net",
|
||||
}
|
||||
"nfs": {
|
||||
"netNamespaceFilePath": "<kubeletRootPath>/plugins/nfs.csi.ceph.com/net",
|
||||
}
|
||||
}
|
||||
]
|
||||
cluster-mapping.json: |-
|
||||
[
|
||||
{
|
||||
"clusterIDMapping": {
|
||||
"clusterID on site1": "clusterID on site2"
|
||||
},
|
||||
"RBDPoolIDMapping": [{
|
||||
"poolID on site1": "poolID on site2"
|
||||
...
|
||||
}],
|
||||
"CephFSFscIDMapping": [{
|
||||
"CephFS FscID on site1": "CephFS FscID on site2"
|
||||
...
|
||||
}]
|
||||
}
|
||||
]
|
||||
metadata:
|
||||
name: ceph-csi-config
|
22
deploy/service-monitor.yaml
Normal file
22
deploy/service-monitor.yaml
Normal file
@ -0,0 +1,22 @@
|
||||
---
|
||||
# This is a servicemonitor that would be used by a prometheus collector to pick
|
||||
# up liveness metrics. The label your prometheus instance is looking for will
|
||||
# need to be supplied and the namespace may need to changed
|
||||
apiVersion: monitoring.coreos.com/v1
|
||||
kind: ServiceMonitor
|
||||
metadata:
|
||||
name: csi-metrics
|
||||
namespace: rook-ceph
|
||||
labels:
|
||||
team: rook
|
||||
spec:
|
||||
namespaceSelector:
|
||||
matchNames:
|
||||
- default
|
||||
selector:
|
||||
matchLabels:
|
||||
app: csi-metrics
|
||||
endpoints:
|
||||
- port: http-metrics
|
||||
path: /metrics
|
||||
interval: 5s
|
Reference in New Issue
Block a user