blob: fd9a3be31e12dbcac0bfc0104261794351313436 [file] [log] [blame]
jpavlik8425d362015-06-09 15:23:27 +02001
Ondrej Smola81d1a192017-08-17 11:13:10 +02002============
3Ceph formula
4============
jpavlik8425d362015-06-09 15:23:27 +02005
Ondrej Smola81d1a192017-08-17 11:13:10 +02006Ceph provides extraordinary data storage scalability. Thousands of client
7hosts or KVMs accessing petabytes to exabytes of data. Each one of your
8applications can use the object, block or file system interfaces to the same
9RADOS cluster simultaneously, which means your Ceph storage system serves as a
10flexible foundation for all of your data storage needs.
jpavlik8425d362015-06-09 15:23:27 +020011
Ondrej Smola81d1a192017-08-17 11:13:10 +020012Use salt-formula-linux for initial disk partitioning.
jpavlik8425d362015-06-09 15:23:27 +020013
14
Tomáš Kukráld2b82972017-08-29 12:45:45 +020015Daemons
16--------
17
18Ceph uses several daemons to handle data and cluster state. Each daemon type requires different computing capacity and hardware optimization.
19
20These daemons are currently supported by formula:
21
22* MON (`ceph.mon`)
23* OSD (`ceph.osd`)
24* RGW (`ceph.radosgw`)
25
26
27Architecture decisions
28-----------------------
29
30Please refer to upstream achritecture documents before designing your cluster. Solid understanding of Ceph principles is essential for making architecture decisions described bellow.
31http://docs.ceph.com/docs/master/architecture/
32
33* Ceph version
34
35There is 3 or 4 stable releases every year and many of nighty/dev release. You should decide which version will be used since the only stable releases are recommended for production. Some of the releases are marked LTS (Long Term Stable) and these releases receive bugfixed for longer period - usually until next LTS version is released.
36
37* Number of MON daemons
38
39Use 1 MON daemon for testing, 3 MONs for smaller production clusters and 5 MONs for very large production cluster. There is no need to have more than 5 MONs in normal environment because there isn't any significant benefit in running more than 5 MONs. Ceph require MONS to form quorum so you need to heve more than 50% of the MONs up and running to have fully operational cluster. Every I/O operation will stop once less than 50% MONs is availabe because they can't form quorum.
40
41* Number of PGs
42
43Placement groups are providing mappping between stored data and OSDs. It is necessary to calculate number of PGs because there should be stored decent amount of PGs on each OSD. Please keep in mind *decreasing number of PGs* isn't possible and *increading* can affect cluster performance.
44
45http://docs.ceph.com/docs/master/rados/operations/placement-groups/
46http://ceph.com/pgcalc/
47
48* Daemon colocation
49
50It is recommended to dedicate nodes for MONs and RWG since colocation can have and influence on cluster operations. Howerver, small clusters can be running MONs on OSD node but it is critical to have enough of resources for MON daemons because they are the most important part of the cluster.
51
52Installing RGW on node with other daemons isn't recommended because RGW daemon usually require a lot of bandwith and it harm cluster health.
53
Tomáš Kukráld2b82972017-08-29 12:45:45 +020054* Store type (Bluestore/Filestore)
55
56Recent version of Ceph support Bluestore as storage backend and backend should be used if available.
57
58http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/
59
Jiri Broulikcc0d7752017-11-18 18:58:21 +010060* Block.db location for Bluestore
61
62There are two ways to setup block.db:
63 * **Colocated** block.db partition is created on the same disk as partition for the data. This setup is easier for installation and it doesn't require any other disk to be used. However, colocated setup is significantly slower than dedicated)
64 * **Dedicate** block.db is placed on different disk than data (or into partition). This setup can deliver much higher performance than colocated but it require to have more disks in servers. Block.db drives should be carefully selected because high I/O and durability is required.
65
66* Block.wal location for Bluestore
67
68There are two ways to setup block.wal - stores just the internal journal (write-ahead log):
69 * **Colocated** block.wal uses free space of the block.db device.
70 * **Dedicate** block.wal is placed on different disk than data (better put into partition as the size can be small) and possibly block.db device. This setup can deliver much higher performance than colocated but it require to have more disks in servers. Block.wal drives should be carefully selected because high I/O and durability is required.
71
72* Journal location for Filestore
73
74There are two ways to setup journal:
75 * **Colocated** journal is created on the same disk as partition for the data. This setup is easier for installation and it doesn't require any other disk to be used. However, colocated setup is significantly slower than dedicated)
76 * **Dedicate** journal is placed on different disk than data (or into partition). This setup can deliver much higher performance than colocated but it require to have more disks in servers. Journal drives should be carefully selected because high I/O and durability is required.
77
Tomáš Kukráld2b82972017-08-29 12:45:45 +020078* Cluster and public network
79
Mateusz Los4dd8c4f2017-12-01 09:53:02 +010080Ceph cluster is accessed using network and thus you need to have decend capacity to handle all the client. There are two networks required for cluster: **public** network and cluster network. Public network is used for client connections and MONs and OSDs are listening on this network. Second network ic called **cluster** networks and this network is used for communication between OSDs.
Tomáš Kukráld2b82972017-08-29 12:45:45 +020081
82Both networks should have dedicated interfaces, bonding interfaces and dedicating vlans on bonded interfaces isn't allowed. Good practise is dedicate more throughput for the cluster network because cluster traffic is more important than client traffic.
83
84* Pool parameters (size, min_size, type)
85
86You should setup each pool according to it's expected usage, at least `min_size` and `size` and pool type should be considered.
87
88* Cluster monitoring
89
90* Hardware
91
92Please refer to upstream hardware recommendation guide for general information about hardware.
93
94Ceph servers are required to fulfil special requirements becauce load generated by Ceph can be diametrically opposed to common load.
95
96http://docs.ceph.com/docs/master/start/hardware-recommendations/
97
98
99Basic management commands
100------------------------------
101
102Cluster
103********
104
105- :code:`ceph health` - check if cluster is healthy (:code:`ceph health detail` can provide more information)
106
107
108.. code-block:: bash
109
110 root@c-01:~# ceph health
111 HEALTH_OK
112
113- :code:`ceph status` - shows basic information about cluster
114
115
116.. code-block:: bash
117
118 root@c-01:~# ceph status
119 cluster e2dc51ae-c5e4-48f0-afc1-9e9e97dfd650
120 health HEALTH_OK
121 monmap e1: 3 mons at {1=192.168.31.201:6789/0,2=192.168.31.202:6789/0,3=192.168.31.203:6789/0}
122 election epoch 38, quorum 0,1,2 1,2,3
123 osdmap e226: 6 osds: 6 up, 6 in
124 pgmap v27916: 400 pgs, 2 pools, 21233 MB data, 5315 objects
125 121 GB used, 10924 GB / 11058 GB avail
126 400 active+clean
127 client io 481 kB/s rd, 132 kB/s wr, 185 op/
128
129MON
130****
131
132http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-mon/
133
134OSD
135****
136
137http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/
138
139- :code:`ceph osd tree` - show all OSDs and it's state
140
141.. code-block:: bash
142
143 root@c-01:~# ceph osd tree
144 ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
145 -4 0 host c-04
146 -1 10.79993 root default
147 -2 3.59998 host c-01
148 0 1.79999 osd.0 up 1.00000 1.00000
149 1 1.79999 osd.1 up 1.00000 1.00000
150 -3 3.59998 host c-02
151 2 1.79999 osd.2 up 1.00000 1.00000
152 3 1.79999 osd.3 up 1.00000 1.00000
153 -5 3.59998 host c-03
154 4 1.79999 osd.4 up 1.00000 1.00000
155 5 1.79999 osd.5 up 1.00000 1.00000
156
157- :code:`ceph osd pools ls` - list of pool
158
159.. code-block:: bash
160
161 root@c-01:~# ceph osd lspools
162 0 rbd,1 test
163
164PG
165***
166
167http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg
168
169- :code:`ceph pg ls` - list placement groups
170
171.. code-block:: bash
172
173 root@c-01:~# ceph pg ls | head -n 4
174 pg_stat objects mip degr misp unf bytes log disklog state state_stamp v reported up up_primary acting acting_primary last_scrub scrub_stamp last_deep_scrub deep_scrub_stamp
175 0.0 11 0 0 0 0 46137344 3044 3044 active+clean 2015-07-02 10:12:40.603692 226'10652 226:1798 [4,2,0] 4 [4,2,0] 4 0'0 2015-07-01 18:38:33.126953 0'0 2015-07-01 18:17:01.904194
176 0.1 7 0 0 0 0 25165936 3026 3026 active+clean 2015-07-02 10:12:40.585833 226'5808 226:1070 [2,4,1] 2 [2,4,1] 2 0'0 2015-07-01 18:38:32.352721 0'0 2015-07-01 18:17:01.904198
177 0.2 18 0 0 0 0 75497472 3039 3039 active+clean 2015-07-02 10:12:39.569630 226'17447 226:3213 [3,1,5] 3 [3,1,5] 3 0'0 2015-07-01 18:38:34.308228 0'0 2015-07-01 18:17:01.904199
178
179- :code:`ceph pg map 1.1` - show mapping between PG and OSD
180
181.. code-block:: bash
182
183 root@c-01:~# ceph pg map 1.1
184 osdmap e226 pg 1.1 (1.1) -> up [5,1,2] acting [5,1,2]
185
186
187
jpavlik8425d362015-06-09 15:23:27 +0200188Sample pillars
189==============
190
Ondrej Smola81d1a192017-08-17 11:13:10 +0200191Common metadata for all nodes/roles
jpavlik8425d362015-06-09 15:23:27 +0200192
193.. code-block:: yaml
194
195 ceph:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200196 common:
Jiri Broulikd5729042017-09-19 20:07:22 +0200197 version: luminous
jpavlik8425d362015-06-09 15:23:27 +0200198 config:
199 global:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200200 param1: value1
201 param2: value1
202 param3: value1
203 pool_section:
204 param1: value2
205 param2: value2
206 param3: value2
207 fsid: a619c5fc-c4ed-4f22-9ed2-66cf2feca23d
208 members:
209 - name: cmn01
210 host: 10.0.0.1
211 - name: cmn02
212 host: 10.0.0.2
213 - name: cmn03
214 host: 10.0.0.3
jpavlik8425d362015-06-09 15:23:27 +0200215 keyring:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200216 admin:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200217 caps:
218 mds: "allow *"
219 mgr: "allow *"
220 mon: "allow *"
221 osd: "allow *"
Jiri Broulikd5729042017-09-19 20:07:22 +0200222 bootstrap-osd:
Jiri Broulikd5729042017-09-19 20:07:22 +0200223 caps:
224 mon: "allow profile bootstrap-osd"
225
jpavlik8425d362015-06-09 15:23:27 +0200226
Ondrej Smola81d1a192017-08-17 11:13:10 +0200227Optional definition for cluster and public networks. Cluster network is used
228for replication. Public network for front-end communication.
jpavlik8425d362015-06-09 15:23:27 +0200229
230.. code-block:: yaml
231
232 ceph:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200233 common:
Jiri Broulikd5729042017-09-19 20:07:22 +0200234 version: luminous
Ondrej Smola81d1a192017-08-17 11:13:10 +0200235 fsid: a619c5fc-c4ed-4f22-9ed2-66cf2feca23d
236 ....
237 public_network: 10.0.0.0/24, 10.1.0.0/24
238 cluster_network: 10.10.0.0/24, 10.11.0.0/24
239
240
241Ceph mon (control) roles
242------------------------
243
244Monitors: A Ceph Monitor maintains maps of the cluster state, including the
245monitor map, the OSD map, the Placement Group (PG) map, and the CRUSH map.
246Ceph maintains a history (called an epoch”) of each state change in the Ceph
247Monitors, Ceph OSD Daemons, and PGs.
248
249.. code-block:: yaml
250
251 ceph:
252 common:
253 config:
254 mon:
255 key: value
jpavlik8425d362015-06-09 15:23:27 +0200256 mon:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200257 enabled: true
jpavlik8425d362015-06-09 15:23:27 +0200258 keyring:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200259 mon:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200260 caps:
261 mon: "allow *"
262 admin:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200263 caps:
264 mds: "allow *"
265 mgr: "allow *"
266 mon: "allow *"
267 osd: "allow *"
jpavlik8425d362015-06-09 15:23:27 +0200268
Ondrej Smola91c83162017-09-12 16:40:02 +0200269Ceph mgr roles
270------------------------
271
272The Ceph Manager daemon (ceph-mgr) runs alongside monitor daemons, to provide additional monitoring and interfaces to external monitoring and management systems. Since the 12.x (luminous) Ceph release, the ceph-mgr daemon is required for normal operations. The ceph-mgr daemon is an optional component in the 11.x (kraken) Ceph release.
273
274By default, the manager daemon requires no additional configuration, beyond ensuring it is running. If there is no mgr daemon running, you will see a health warning to that effect, and some of the other information in the output of ceph status will be missing or stale until a mgr is started.
275
276
277.. code-block:: yaml
278
279 ceph:
280 mgr:
281 enabled: true
282 dashboard:
283 enabled: true
284 host: 10.103.255.252
285 port: 7000
286
Ondrej Smola81d1a192017-08-17 11:13:10 +0200287
288Ceph OSD (storage) roles
289------------------------
jpavlik8425d362015-06-09 15:23:27 +0200290
291.. code-block:: yaml
292
293 ceph:
Ondrej Smola81d1a192017-08-17 11:13:10 +0200294 common:
Jiri Broulikec62dec2017-10-10 13:45:15 +0200295 version: luminous
296 fsid: a619c5fc-c4ed-4f22-9ed2-66cf2feca23d
297 public_network: 10.0.0.0/24, 10.1.0.0/24
298 cluster_network: 10.10.0.0/24, 10.11.0.0/24
299 keyring:
300 bootstrap-osd:
301 caps:
302 mon: "allow profile bootstrap-osd"
303 ....
Ondrej Smola81d1a192017-08-17 11:13:10 +0200304 osd:
305 enabled: true
Jiri Broulikec62dec2017-10-10 13:45:15 +0200306 crush_parent: rack01
307 journal_size: 20480 (20G)
308 bluestore_block_db_size: 10073741824 (10G)
309 bluestore_block_wal_size: 10073741824 (10G)
Jiri Broulikd5729042017-09-19 20:07:22 +0200310 bluestore_block_size: 807374182400 (800G)
311 backend:
312 filestore:
313 disks:
314 - dev: /dev/sdm
315 enabled: false
Jiri Broulikd5729042017-09-19 20:07:22 +0200316 journal: /dev/ssd
Jiri Broulik8870b872018-01-24 18:04:25 +0100317 journal_partition: 5
318 data_partition: 6
319 lockbox_partition: 7
320 data_partition_size: 12000 (MB)
Jiri Broulikd5729042017-09-19 20:07:22 +0200321 class: bestssd
Jiri Broulik8870b872018-01-24 18:04:25 +0100322 weight: 1.666
Jiri Broulik58ff84b2017-11-21 14:23:51 +0100323 dmcrypt: true
Jiri Broulik8870b872018-01-24 18:04:25 +0100324 journal_dmcrypt: false
325 - dev: /dev/sdf
326 journal: /dev/ssd
327 journal_dmcrypt: true
328 class: bestssd
329 weight: 1.666
Jiri Broulikd5729042017-09-19 20:07:22 +0200330 - dev: /dev/sdl
Jiri Broulikd5729042017-09-19 20:07:22 +0200331 journal: /dev/ssd
Jiri Broulikd5729042017-09-19 20:07:22 +0200332 class: bestssd
Jiri Broulik8870b872018-01-24 18:04:25 +0100333 weight: 1.666
Jiri Broulikd5729042017-09-19 20:07:22 +0200334 bluestore:
335 disks:
336 - dev: /dev/sdb
Jiri Broulik8870b872018-01-24 18:04:25 +0100337 - dev: /dev/sdf
338 block_db: /dev/ssd
339 block_wal: /dev/ssd
340 block_db_dmcrypt: true
341 block_wal_dmcrypt: true
Jiri Broulikd5729042017-09-19 20:07:22 +0200342 - dev: /dev/sdc
343 block_db: /dev/ssd
344 block_wal: /dev/ssd
Jiri Broulik8870b872018-01-24 18:04:25 +0100345 data_partition: 1
346 block_partition: 2
347 lockbox_partition: 5
348 block_db_partition: 3
349 block_wal_partition: 4
Jiri Broulikc2be93b2017-10-03 14:20:00 +0200350 class: ssd
351 weight: 1.666
Jiri Broulik58ff84b2017-11-21 14:23:51 +0100352 dmcrypt: true
Jiri Broulik8870b872018-01-24 18:04:25 +0100353 block_db_dmcrypt: false
354 block_wal_dmcrypt: false
Jiri Broulikd5729042017-09-19 20:07:22 +0200355 - dev: /dev/sdd
356 enabled: false
jpavlik8425d362015-06-09 15:23:27 +0200357
Ondrej Smola81d1a192017-08-17 11:13:10 +0200358
Jiri Broulikc2be93b2017-10-03 14:20:00 +0200359Ceph client roles - ...Deprecated - use ceph:common instead
360--------------------------------------------------------
Ondrej Smola81d1a192017-08-17 11:13:10 +0200361
362Simple ceph client service
Simon Pasquierf8e6f9e2017-07-03 10:15:20 +0200363
364.. code-block:: yaml
365
366 ceph:
367 client:
368 config:
369 global:
370 mon initial members: ceph1,ceph2,ceph3
371 mon host: 10.103.255.252:6789,10.103.255.253:6789,10.103.255.254:6789
372 keyring:
373 monitoring:
374 key: 00000000000000000000000000000000000000==
Ondrej Smola81d1a192017-08-17 11:13:10 +0200375
376At OpenStack control settings are usually located at cinder-volume or glance-
377registry services.
378
379.. code-block:: yaml
380
381 ceph:
382 client:
383 config:
384 global:
385 fsid: 00000000-0000-0000-0000-000000000000
386 mon initial members: ceph1,ceph2,ceph3
387 mon host: 10.103.255.252:6789,10.103.255.253:6789,10.103.255.254:6789
388 osd_fs_mkfs_arguments_xfs:
389 osd_fs_mount_options_xfs: rw,noatime
390 network public: 10.0.0.0/24
391 network cluster: 10.0.0.0/24
392 osd_fs_type: xfs
393 osd:
394 osd journal size: 7500
395 filestore xattr use omap: true
396 mon:
397 mon debug dump transactions: false
398 keyring:
399 cinder:
400 key: 00000000000000000000000000000000000000==
401 glance:
402 key: 00000000000000000000000000000000000000==
403
404
405Ceph gateway
406------------
407
408Rados gateway with keystone v2 auth backend
409
410.. code-block:: yaml
411
412 ceph:
413 radosgw:
414 enabled: true
415 hostname: gw.ceph.lab
416 bind:
417 address: 10.10.10.1
418 port: 8080
419 identity:
420 engine: keystone
421 api_version: 2
422 host: 10.10.10.100
423 port: 5000
424 user: admin
425 password: password
426 tenant: admin
427
428Rados gateway with keystone v3 auth backend
429
430.. code-block:: yaml
431
432 ceph:
433 radosgw:
434 enabled: true
435 hostname: gw.ceph.lab
436 bind:
437 address: 10.10.10.1
438 port: 8080
439 identity:
440 engine: keystone
441 api_version: 3
442 host: 10.10.10.100
443 port: 5000
444 user: admin
445 password: password
446 project: admin
447 domain: default
448
449
450Ceph setup role
451---------------
452
453Replicated ceph storage pool
454
455.. code-block:: yaml
456
457 ceph:
458 setup:
459 pool:
460 replicated_pool:
461 pg_num: 256
462 pgp_num: 256
463 type: replicated
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200464 crush_rule: sata
465 application: rbd
Ondrej Smola81d1a192017-08-17 11:13:10 +0200466
Jiri Broulikeaf41472017-10-18 09:56:33 +0200467 .. note:: For Kraken and earlier releases please specify crush_rule as a ruleset number.
468 For Kraken and earlier releases application param is not needed.
469
Ondrej Smola81d1a192017-08-17 11:13:10 +0200470Erasure ceph storage pool
471
472.. code-block:: yaml
473
474 ceph:
475 setup:
476 pool:
477 erasure_pool:
478 pg_num: 256
479 pgp_num: 256
480 type: erasure
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200481 crush_rule: ssd
482 application: rbd
Ondrej Smola81d1a192017-08-17 11:13:10 +0200483
Jiri Broulikd68e33a2017-10-24 10:54:43 +0200484
Jiri Broulike4ba9f62017-11-08 11:33:00 +0100485Inline compression for Bluestore backend
486
487.. code-block:: yaml
488
489 ceph:
490 setup:
491 pool:
492 volumes:
493 pg_num: 256
494 pgp_num: 256
495 type: replicated
496 crush_rule: hdd
497 application: rbd
498 compression_algorithm: snappy
499 compression_mode: aggressive
500 compression_required_ratio: .875
501 ...
502
503
Jiri Broulikd68e33a2017-10-24 10:54:43 +0200504Ceph manage keyring keys
505------------------------
506
507Keyrings are dynamically generated unless specified by the following pillar.
508
509.. code-block:: yaml
510
511 ceph:
512 common:
513 manage_keyring: true
514 keyring:
515 glance:
516 name: images
517 key: AACf3ulZFFPNDxAAd2DWds3aEkHh4IklZVgIaQ==
518 caps:
519 mon: "allow r"
520 osd: "allow class-read object_prefix rdb_children, allow rwx pool=images"
521
522
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200523Generate CRUSH map - Recommended way
524-----------------------------------
Tomáš Kukrál363d37d2017-08-17 13:40:20 +0200525
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200526It is required to define the `type` for crush buckets and these types must start with `root` (top) and end with `host`. OSD daemons will be assigned to hosts according to it's hostname. Weight of the buckets will be calculated according to weight of it's children.
527
528If the pools that are in use have size of 3 it is best to have 3 children of a specific type in the root CRUSH tree to replicate objects across (Specified in rule steps by 'type region').
Tomáš Kukrál363d37d2017-08-17 13:40:20 +0200529
530.. code-block:: yaml
531
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200532 ceph:
533 setup:
534 crush:
535 enabled: True
536 tunables:
537 choose_total_tries: 50
538 choose_local_tries: 0
539 choose_local_fallback_tries: 0
540 chooseleaf_descend_once: 1
541 chooseleaf_vary_r: 1
542 chooseleaf_stable: 1
543 straw_calc_version: 1
544 allowed_bucket_algs: 54
545 type:
546 - root
547 - region
548 - rack
549 - host
Jiri Broulikeaf41472017-10-18 09:56:33 +0200550 - osd
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200551 root:
552 - name: root-ssd
553 - name: root-sata
554 region:
555 - name: eu-1
556 parent: root-sata
557 - name: eu-2
558 parent: root-sata
559 - name: eu-3
560 parent: root-ssd
561 - name: us-1
562 parent: root-sata
563 rack:
564 - name: rack01
565 parent: eu-1
566 - name: rack02
567 parent: eu-2
568 - name: rack03
569 parent: us-1
570 rule:
571 sata:
572 ruleset: 0
573 type: replicated
574 min_size: 1
575 max_size: 10
576 steps:
577 - take take root-ssd
578 - chooseleaf firstn 0 type region
579 - emit
580 ssd:
581 ruleset: 1
582 type: replicated
583 min_size: 1
584 max_size: 10
585 steps:
586 - take take root-sata
587 - chooseleaf firstn 0 type region
588 - emit
589
590
591Generate CRUSH map - Alternative way
592------------------------------------
593
594It's necessary to create per OSD pillar.
595
596.. code-block:: yaml
597
598 ceph:
599 osd:
600 crush:
601 - type: root
602 name: root1
603 - type: region
604 name: eu-1
605 - type: rack
606 name: rack01
607 - type: host
608 name: osd001
609
Jiri Broulik8870b872018-01-24 18:04:25 +0100610Add OSDs with specific weight
611-----------------------------
612
613Add OSD device(s) with initial weight set specifically to certain value.
614
615.. code-block:: yaml
616
617 ceph:
618 osd:
619 crush_initial_weight: 0
620
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200621
622Apply CRUSH map
623---------------
624
625Before you apply CRUSH map please make sure that settings in generated file in /etc/ceph/crushmap are correct.
626
627.. code-block:: yaml
628
629 ceph:
630 setup:
631 crush:
632 enforce: true
633 pool:
634 images:
635 crush_rule: sata
636 application: rbd
637 volumes:
638 crush_rule: sata
639 application: rbd
640 vms:
641 crush_rule: ssd
642 application: rbd
643
Jiri Broulikeaf41472017-10-18 09:56:33 +0200644 .. note:: For Kraken and earlier releases please specify crush_rule as a ruleset number.
645 For Kraken and earlier releases application param is not needed.
646
Jiri Broulik97af8ab2017-10-12 14:32:51 +0200647
648Persist CRUSH map
649--------------------
650
651After the CRUSH map is applied to Ceph it's recommended to persist the same settings even after OSD reboots.
652
653.. code-block:: yaml
654
655 ceph:
656 osd:
657 crush_update: false
658
Ondrej Smola81d1a192017-08-17 11:13:10 +0200659
660Ceph monitoring
661---------------
662
Jiri Broulik44574072017-11-14 12:27:39 +0100663By default monitoring is setup to collect information from MON and OSD nodes. To change the default values add the following pillar to MON nodes.
Ondrej Smola81d1a192017-08-17 11:13:10 +0200664
665.. code-block:: yaml
666
667 ceph:
Simon Pasquierf8e6f9e2017-07-03 10:15:20 +0200668 monitoring:
Jiri Broulik44574072017-11-14 12:27:39 +0100669 space_used_warning_threshold: 0.75
670 space_used_critical_threshold: 0.85
671 apply_latency_threshold: 0.007
672 commit_latency_threshold: 0.7
673 pool_space_used_utilization_warning_threshold: 0.75
674 pool_space_used_critical_threshold: 0.85
675 pool_write_ops_threshold: 200
676 pool_write_bytes_threshold: 70000000
677 pool_read_bytes_threshold: 70000000
678 pool_read_ops_threshold: 1000
Simon Pasquierf8e6f9e2017-07-03 10:15:20 +0200679
Mateusz Los4dd8c4f2017-12-01 09:53:02 +0100680Ceph monitor backups
681--------------------
682
683Backup client with ssh/rsync remote host
684
685.. code-block:: yaml
686
687 ceph:
688 backup:
689 client:
690 enabled: true
691 full_backups_to_keep: 3
692 hours_before_full: 24
693 target:
694 host: cfg01
695
696
697Backup client with local backup only
698
699.. code-block:: yaml
700
701 ceph:
702 backup:
703 client:
704 enabled: true
705 full_backups_to_keep: 3
706 hours_before_full: 24
707
708Backup server rsync
709
710.. code-block:: yaml
711
712 ceph:
713 backup:
714 server:
715 enabled: true
716 hours_before_full: 24
717 full_backups_to_keep: 5
718 key:
719 ceph_pub_key:
720 enabled: true
721 key: ssh_rsa
722
723
Simon Pasquierf8e6f9e2017-07-03 10:15:20 +0200724
Ondrej Smola81d1a192017-08-17 11:13:10 +0200725More information
726================
jpavlik8425d362015-06-09 15:23:27 +0200727
728* https://github.com/cloud-ee/ceph-salt-formula
729* http://ceph.com/ceph-storage/
jan kaufman4f7757b2015-06-12 10:49:00 +0200730* http://ceph.com/docs/master/start/intro/
Filip Pytloun32841d72017-02-02 13:02:03 +0100731
Ondrej Smola81d1a192017-08-17 11:13:10 +0200732
733Documentation and bugs
Filip Pytloun32841d72017-02-02 13:02:03 +0100734======================
735
736To learn how to install and update salt-formulas, consult the documentation
737available online at:
738
739 http://salt-formulas.readthedocs.io/
740
741In the unfortunate event that bugs are discovered, they should be reported to
742the appropriate issue tracker. Use Github issue tracker for specific salt
743formula:
744
745 https://github.com/salt-formulas/salt-formula-ceph/issues
746
747For feature requests, bug reports or blueprints affecting entire ecosystem,
748use Launchpad salt-formulas project:
749
750 https://launchpad.net/salt-formulas
751
752You can also join salt-formulas-users team and subscribe to mailing list:
753
754 https://launchpad.net/~salt-formulas-users
755
756Developers wishing to work on the salt-formulas projects should always base
757their work on master branch and submit pull request against specific formula.
758
759 https://github.com/salt-formulas/salt-formula-ceph
760
761Any questions or feedback is always welcome so feel free to join our IRC
762channel:
763
764 #salt-formulas @ irc.freenode.net