ceph cluster sample configure file in src

简介:
ceph集群的配置文件模板如下 : 

[root@localhost src]# cat sample.ceph.conf 
##
# Sample ceph ceph.conf file.
##
# This file defines cluster membership, the various locations
# that Ceph stores data, and any other runtime options.

# If a 'host' is defined for a daemon, the init.d start/stop script will
# verify that it matches the hostname (or else ignore it).  If it is
# not defined, it is assumed that the daemon is intended to start on
# the current host (e.g., in a setup with a startup.conf on each
# node).

配置文件中支持的变量 : 
## Metavariables
# $cluster    ; Expands to the Ceph Storage Cluster name. Useful
#             ; when running multiple Ceph Storage Clusters
#             ; on the same hardware.
#             ; Example: /etc/ceph/$cluster.keyring
#             ; (Default: ceph)
#
# $type       ; Expands to one of mds, osd, or mon, depending on
#             ; the type of the instant daemon.
#             ; Example: /var/lib/ceph/$type
#
# $id         ; Expands to the daemon identifier. For osd.0, this
#             ; would be 0; for mds.a, it would be a.
#             ; Example: /var/lib/ceph/$type/$cluster-$id
#
# $host       ; Expands to the host name of the instant daemon.
#
# $name       ; Expands to $type.$id.
#             ; Example: /var/run/ceph/$cluster-$name.asok
全局配置章节
[global]
### http://ceph.com/docs/master/rados/configuration/general-config-ref/

    ;fsid                       = {UUID}    # use `uuidgen` to generate your own UUID
    ;public network             = 192.168.0.0/24
    ;cluster network            = 192.168.0.0/24

    # Each running Ceph daemon has a running process identifier (PID) file.
    # The PID file is generated upon start-up.
    # Type: String (optional)
    # (Default: N/A). The default path is /var/run/$cluster/$name.pid.
    pid file                   = /var/run/ceph/$name.pid

    # If set, when the Ceph Storage Cluster starts, Ceph sets the max open fds
    # at the OS level (i.e., the max # of file descriptors).
    # It helps prevents Ceph OSD Daemons from running out of file descriptors.
    # Type: 64-bit Integer (optional)
    # (Default: 0)
    ;max open files             = 131072


### http://ceph.com/docs/master/rados/operations/authentication
### http://ceph.com/docs/master/rados/configuration/auth-config-ref/

    # If enabled, the Ceph Storage Cluster daemons (i.e., ceph-mon, ceph-osd,
    # and ceph-mds) must authenticate with each other.
    # Type: String (optional); Valid settings are "cephx" or "none".
    # (Default: cephx)
    auth cluster required      = cephx

    # If enabled, the Ceph Storage Cluster daemons require Ceph Clients to
    # authenticate with the Ceph Storage Cluster in order to access Ceph
    # services.
    # Type: String (optional); Valid settings are "cephx" or "none".
    # (Default: cephx)
    auth service required      = cephx

    # If enabled, the Ceph Client requires the Ceph Storage Cluster to
    # authenticate with the Ceph Client.
    # Type: String (optional); Valid settings are "cephx" or "none".
    # (Default: cephx)
    auth client required       = cephx

    # If set to true, Ceph requires signatures on all message traffic between
    # the Ceph Client and the Ceph Storage Cluster, and between daemons
    # comprising the Ceph Storage Cluster.
    # Type: Boolean (optional)
    # (Default: false)
    ;cephx require signatures   = true

    # kernel RBD client do not support authentication yet:
    cephx cluster require signatures = true
    cephx service require signatures = false

    # The path to the keyring file.
    # Type: String (optional)
    # Default: /etc/ceph/$cluster.$name.keyring,/etc/ceph/$cluster.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin
    ;keyring                  = /etc/ceph/$cluster.$name.keyring


### http://ceph.com/docs/master/rados/configuration/pool-pg-config-ref/


    ## Replication level, number of data copies.
    # Type: 32-bit Integer
    # (Default: 3)
    ;osd pool default size      = 3

    ## Replication level in degraded state, less than 'osd pool default size' value.
    # Sets the minimum number of written replicas for objects in the
    # pool in order to acknowledge a write operation to the client. If
    # minimum is not met, Ceph will not acknowledge the write to the
    # client. This setting ensures a minimum number of replicas when
    # operating in degraded mode.
    # Type: 32-bit Integer
    # (Default: 0), which means no particular minimum. If 0, minimum is size - (size / 2).
    ;osd pool default min size  = 2

    ## Ensure you have a realistic number of placement groups. We recommend
    ## approximately 100 per OSD. E.g., total number of OSDs multiplied by 100
    ## divided by the number of replicas (i.e., osd pool default size). So for
    ## 10 OSDs and osd pool default size = 3, we'd recommend approximately
    ## (100 * 10) / 3 = 333

    # Description: The default number of placement groups for a pool. The
    #              default value is the same as pg_num with mkpool.
    # Type: 32-bit Integer
    # (Default: 8)
    ;osd pool default pg num    = 128

    # Description: The default number of placement groups for placement for a
    #              pool. The default value is the same as pgp_num with mkpool.
    #              PG and PGP should be equal (for now).
    # Type: 32-bit Integer
    # (Default: 8)
    ;osd pool default pgp num   = 128

    # The default CRUSH ruleset to use when creating a pool
    # Type: 32-bit Integer
    # (Default: 0)
    ;osd pool default crush rule = 0

    # The bucket type to use for chooseleaf in a CRUSH rule.
    # Uses ordinal rank rather than name.
    # Type: 32-bit Integer
    # (Default: 1) Typically a host containing one or more Ceph OSD Daemons.
    ;osd crush chooseleaf type = 1


### http://ceph.com/docs/master/rados/troubleshooting/log-and-debug/

    # The location of the logging file for your cluster.
    # Type: String
    # Required: No
    # Default: /var/log/ceph/$cluster-$name.log
    ;log file                   = /var/log/ceph/$cluster-$name.log

    # Determines if logging messages should appear in syslog.
    # Type: Boolean
    # Required: No
    # (Default: false)
    ;log to syslog              = true


### http://ceph.com/docs/master/rados/configuration/ms-ref/

    # Enable if you want your daemons to bind to IPv6 address instead of
    # IPv4 ones. (Not required if you specify a daemon or cluster IP.)
    # Type: Boolean
    # (Default: false)
    ;ms bind ipv6               = true

监控配置章节
##################
## Monitors
## You need at least one. You need at least three if you want to
## tolerate any node failures. Always create an odd number.
[mon]
### http://ceph.com/docs/master/rados/configuration/mon-config-ref/
### http://ceph.com/docs/master/rados/configuration/mon-osd-interaction/

    # The IDs of initial monitors in a cluster during startup.
    # If specified, Ceph requires an odd number of monitors to form an
    # initial quorum (e.g., 3).
    # Type: String
    # (Default: None)
    ;mon initial members        = mycephhost

    ;mon host                   = cephhost01,cephhost02
    ;mon addr                   = 192.168.0.101,192.168.0.102

    # The monitor's data location
    # Default: /var/lib/ceph/mon/$cluster-$id
    ;mon data                   = /var/lib/ceph/mon/$name

    # The clock drift in seconds allowed between monitors.
    # Type: Float
    # (Default: .050)
    ;mon clock drift allowed    = .15

    # Exponential backoff for clock drift warnings
    # Type: Float
    # (Default: 5)
    ;mon clock drift warn backoff = 30    # Tell the monitor to backoff from this warning for 30 seconds

    # The percentage of disk space used before an OSD is considered full.
    # Type: Float
    # (Default: .95)
    ;mon osd full ratio         = .95

    # The percentage of disk space used before an OSD is considered nearfull.
    # Type: Float
    # (Default: .85)
    ;mon osd nearfull ratio     = .85

    # The number of seconds Ceph waits before marking a Ceph OSD
    # Daemon "down" and "out" if it doesn't respond.
    # Type: 32-bit Integer
    # (Default: 300)
    ;mon osd down out interval  = 300

    # The grace period in seconds before declaring unresponsive Ceph OSD
    # Daemons "down".
    # Type: 32-bit Integer
    # (Default: 900)
    ;mon osd report timeout          = 300

### http://ceph.com/docs/master/rados/troubleshooting/log-and-debug/

    # logging, for debugging monitor crashes, in order of
    # their likelihood of being helpful :)
    ;debug ms                   = 1
    ;debug mon                  = 20
    ;debug paxos                = 20
    ;debug auth                 = 20


;[mon.alpha]
;    host                       = alpha
;    mon addr                   = 192.168.0.10:6789

;[mon.beta]
;    host                       = beta
;    mon addr                   = 192.168.0.11:6789

;[mon.gamma]
;    host                       = gamma
;    mon addr                   = 192.168.0.12:6789

元数据服务章节(mds)
##################
## Metadata servers
# You must deploy at least one metadata server to use CephFS. There is
# experimental support for running multiple metadata servers. Do not run
# multiple metadata servers in production.
[mds]
### http://ceph.com/docs/master/cephfs/mds-config-ref/

    # where the mds keeps it's secret encryption keys
    ;keyring                    = /var/lib/ceph/mds/$name/keyring

    # Determines whether a 'ceph-mds' daemon should poll and
    # replay the log of an active MDS (hot standby).
    # Type:  Boolean
    # (Default: false)
    ;mds standby replay          = true

    ; mds logging to debug issues.
    ;debug ms                   = 1
    ;debug mds                  = 20
    ;debug journaler            = 20

    # The number of inodes to cache.
    # Type: 32-bit Integer
    # (Default: 100000)
    ;mds cache size             = 250000

    ;mds mem max                = 1048576     # KB

;[mds.alpha]
;    host                       = alpha

;[mds.beta]
;    host                       = beta

osd章节
##################
## osd
# You need at least one.  Two or more if you want data to be replicated.
# Define as many as you like.
[osd]
### http://ceph.com/docs/master/rados/configuration/osd-config-ref/

    # The path to the OSDs data.
    # You must create the directory when deploying Ceph.
    # You should mount a drive for OSD data at this mount point.
    # We do not recommend changing the default.
    # Type: String
    # Default: /var/lib/ceph/osd/$cluster-$id
    ;osd data                     = /var/lib/ceph/osd/$name

    ## You can change the number of recovery operations to speed up recovery
    ## or slow it down if your machines can't handle it

    # The number of active recovery requests per OSD at one time.
    # More requests will accelerate recovery, but the requests
    # places an increased load on the cluster.
    # Type: 32-bit Integer
    # (Default: 5)
    ;osd recovery max active      = 3

    # The maximum number of backfills allowed to or from a single OSD.
    # Type: 64-bit Integer
    # (Default: 10)
    ;osd max backfills            = 5

    # The maximum number of simultaneous scrub operations for a Ceph OSD Daemon.
    # Type: 32-bit Int
    # (Default: 1)
    ;osd max scrubs               = 2

    # You may add settings for ceph-deploy so that it will create and mount
    # the correct type of file system. Remove the comment `#` character for
    # the following settings and replace the values in parenthesis
    # with appropriate values, or leave the following settings commented
    # out to accept the default values.

    #osd mkfs type = {fs-type}
    #osd mkfs options {fs-type}   = {mkfs options}   # default for xfs is "-f"
    #osd mount options {fs-type}  = {mount options} # default mount option is "rw, noatime"
    ;osd mkfs type                = btrfs
    ;osd mount options btrfs      = noatime,nodiratime

    ## Ideally, make this a separate disk or partition.  A few
    ## hundred MB should be enough; more if you have fast or many
    ## disks.  You can use a file under the osd data dir if need be
    ## (e.g. /data/$name/journal), but it will be slower than a
    ## separate disk or partition.
    # The path to the OSD's journal. This may be a path to a file or a block
    # device (such as a partition of an SSD). If it is a file, you must
    # create the directory to contain it.
    # We recommend using a drive separate from the osd data drive.
    # Type: String
    # Default: /var/lib/ceph/osd/$cluster-$id/journal
    ;osd journal                  = /var/lib/ceph/osd/$name/journal

    # Check log files for corruption. Can be computationally expensive.
    # Type: Boolean
    # (Default: false)
    ;osd check for log corruption = true

### http://ceph.com/docs/master/rados/configuration/journal-ref/

    # The size of the journal in megabytes. If this is 0,
    # and the journal is a block device, the entire block device is used.
    # Since v0.54, this is ignored if the journal is a block device,
    # and the entire block device is used.
    # Type: 32-bit Integer
    # (Default: 5120)
    # Recommended: Begin with 1GB. Should be at least twice the product
    # of the expected speed multiplied by "filestore max sync interval".
    ;osd journal size             = 2048     ; journal size, in megabytes

    ## If you want to run the journal on a tmpfs, disable DirectIO
    # Enables direct i/o to the journal.
    # Requires "journal block align" set to "true".
    # Type: Boolean
    # Required: Yes when using aio.
    # (Default: true)
    ;journal dio                  = false

    # osd logging to debug osd issues, in order of likelihood of being helpful
    ;debug ms                     = 1
    ;debug osd                    = 20
    ;debug filestore              = 20
    ;debug journal                = 20

### http://ceph.com/docs/master/rados/configuration/filestore-config-ref/

    # The maximum interval in seconds for synchronizing the filestore.
    # Type: Double (optional)
    # (Default: 5)
    ;filestore max sync interval = 5

    # Enable snapshots for a btrfs filestore.
    # Type: Boolean
    # Required: No. Only used for btrfs.
    # (Default: true)
    ;filestore btrfs snap        = false

    # Enables the filestore flusher.
    # Type: Boolean
    # Required: No
    # (Default: false)
    ;filestore flusher            = true

    # Defines the maximum number of in progress operations the file store
    # accepts before blocking on queuing new operations.
    # Type: Integer
    # Required: No. Minimal impact on performance.
    # (Default: 500)
    ;filestore queue max ops      = 500

    ## Filestore and OSD settings can be tweak to achieve better performance

### http://ceph.com/docs/master/rados/configuration/filestore-config-ref/#misc

    # Min number of files in a subdir before merging into parent NOTE: A negative value means to disable subdir merging
    # Type: Integer
    # Required: No
    # Default:  10
    ;filestore merge threshold    = 10

    # filestore_split_multiple * abs(filestore_merge_threshold) * 16 is the maximum number of files in a subdirectory before splitting into child directories.
    # Type: Integer
    # Required: No
    # Default:  2
    ;filestore split multiple     = 2

    # The number of filesystem operation threads that execute in parallel.
    # Type: Integer
    # Required: No
    # Default:  2
    ;filestore op threads         = 4

    # The number of threads to service Ceph OSD Daemon operations. Set to 0 to disable it. Increasing the number may increase the request processing rate.
    # Type: 32-bit Integer
    # Default:  2
    ;osd op threads               = 2

    ## CRUSH

    # By default OSDs update their details (location, weight and root) on the CRUSH map during startup
    # Type: Boolean
    # Required: No;
    # (Default: true)
    ;osd crush update on start    = false

;[osd.0]
;    host                         = delta

;[osd.1]
;    host                         = epsilon

;[osd.2]
;    host                         = zeta

;[osd.3]
;    host                         = eta

块设备使用配置, 如缓存
##################
## client settings
[client]

### http://ceph.com/docs/master/rbd/rbd-config-ref/

    # Enable caching for RADOS Block Device (RBD).
    # Type: Boolean
    # Required: No
    # (Default: false)
    rbd cache                  = true

    # The RBD cache size in bytes.
    # Type: 64-bit Integer
    # Required: No
    # (Default: 32 MiB)
    ;rbd cache size            = 33554432

    # The dirty limit in bytes at which the cache triggers write-back.
    # If 0, uses write-through caching.
    # Type: 64-bit Integer
    # Required: No
    # Constraint: Must be less than rbd cache size.
    # (Default: 24 MiB)
    ;rbd cache max dirty       = 25165824

    # The dirty target before the cache begins writing data to the data storage.
    # Does not block writes to the cache.
    # Type: 64-bit Integer
    # Required: No
    # Constraint: Must be less than rbd cache max dirty.
    # (Default: 16 MiB)
    ;rbd cache target dirty    = 16777216

    # The number of seconds dirty data is in the cache before writeback starts.
    # Type: Float
    # Required: No
    # (Default: 1.0)
    ;rbd cache max dirty age   = 1.0

    # Start out in write-through mode, and switch to write-back after the
    # first flush request is received. Enabling this is a conservative but
    # safe setting in case VMs running on rbd are too old to send flushes,
    # like the virtio driver in Linux before 2.6.32.
    # Type: Boolean
    # Required: No
    # (Default: false)
    ;rbd cache writethrough until flush = false

对象存储客户端设置
##################
## radosgw client settings
[client.radosgw.gateway]

### http://ceph.com/docs/master/radosgw/config-ref/

    # Sets the location of the data files for Ceph Object Gateway.
    # You must create the directory when deploying Ceph.
    # We do not recommend changing the default.
    # Type: String
    # Default: /var/lib/ceph/radosgw/$cluster-$id
    ;rgw data                     = /var/lib/ceph/radosgw/$name

    # Client's hostname
    ;host                         =  ceph-radosgw

    # where the radosgw keeps it's secret encryption keys
    ;keyring                      = /etc/ceph/ceph.client.radosgw.keyring

    # FastCgiExternalServer uses this socket.
    # If you do not specify a socket path, Ceph Object Gateway will not run as an external server.
    # The path you specify here must be the same as the path specified in the rgw.conf file.
    # Type: String
    # Default: None
    ;rgw socket path              = /var/run/ceph/ceph.radosgw.gateway.fastcgi.sock

    # The location of the logging file for your radosgw.
    # Type: String
    # Required: No
    # Default: /var/log/ceph/$cluster-$name.log
    ;log file                     = /var/log/ceph/client.radosgw.gateway.log

    # Enable 100-continue if it is operational.
    # Type: Boolean
    # Default:  true
    ;rgw print continue           = false

    # The DNS name of the served domain.
    # Type: String
    # Default:  None
    ;rgw dns name                 = radosgw.ceph.internal

[参考]
1. src/sample.ceph.conf
相关文章
|
18天前
|
Kubernetes 容器
Failed at step CHDIR spawning /opt/kubernetes/bin/etcd: No such file or directory
Failed at step CHDIR spawning /opt/kubernetes/bin/etcd: No such file or directory
21 0
|
3月前
|
Linux iOS开发 MacOS
pnpm全局安装报错:Run “pnpm setup“ to create it automatically, or set the global-bin-dir setting, or the PN
pnpm全局安装报错:Run “pnpm setup“ to create it automatically, or set the global-bin-dir setting, or the PN
490 0
|
应用服务中间件
./configure *: No such file or directory
今天在CentOS下安装Nginx时,遇到了类似以下的问题: ./configure auto/unix: No such file or directory ./configure auto/options: No such file or directory 之前在另一台服务器上安装的时候也没有遇到此问题, 安装过程参考:CentOS6.5下Nginx1.7.4安装记录 于是网上搜索解决方案,但是搜索了很久也没有找到解决方案, 当我使用“mkdir auto/unix”之后,还是提示类似的错误,但目录名又变了。
3057 0
|
18天前
|
Kubernetes 网络协议 Perl
k8s Failed to create pod sandbox: open /run/systemd/resolve/resolv.conf: no such file or directory
k8s Failed to create pod sandbox: open /run/systemd/resolve/resolv.conf: no such file or directory
20 0
|
5月前
|
Ubuntu
./autogen.sh: 5: ./autogen.sh: autoreconf: not found
./autogen.sh: 5: ./autogen.sh: autoreconf: not found
|
7月前
cancal报错 config dir not found
cancal报错 config dir not found
./autogen.sh: 3: gtkdocize: not found
./autogen.sh: 3: gtkdocize: not found
75 0
|
分布式计算 安全 Hadoop
hdfs mkdir报错Cannot create directory /usr. Name node is in safe mode.
hdfs mkdir报错Cannot create directory /usr. Name node is in safe mode.
241 0
That command depends on command in Target 'XXX': script phase “[CP] Copy Pods Resourc 报错解决方法
That command depends on command in Target 'XXX': script phase “[CP] Copy Pods Resourc 报错解决方法
147 0
|
iOS开发 MacOS
如何解决npm命令报错Missing write access to /usr/local/lib/node_modules
如何解决npm命令报错Missing write access to /usr/local/lib/node_modules
如何解决npm命令报错Missing write access to /usr/local/lib/node_modules