-
Notifications
You must be signed in to change notification settings - Fork 312
/
topology.example.yaml
398 lines (387 loc) · 15.3 KB
/
topology.example.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
# # Global variables are applied to all deployments and used as the default value of
# # the deployments if a specific deployment value is missing.
global:
# # The user who runs the tidb cluster.
user: "tidb"
# # group is used to specify the group name the user belong to,if it's not the same as user.
# group: "tidb"
# # SSH port of servers in the managed cluster.
# # systemd_mode is used to select whether to use sudo permissions. When its value is set to user, there is no need to add global.user to sudoers. The default value is system.
# systemd_mode: "system"
ssh_port: 22
# # Storage directory for cluster deployment files, startup scripts, and configuration files.
deploy_dir: "/tidb-deploy"
# # TiDB Cluster data storage directory
data_dir: "/tidb-data"
# # default listen_host for all components
listen_host: 0.0.0.0
# # Supported values: "amd64", "arm64" (default: "amd64")
arch: "amd64"
# # Resource Control is used to limit the resource of an instance.
# # See: https://www.freedesktop.org/software/systemd/man/systemd.resource-control.html
# # Supports using instance-level `resource_control` to override global `resource_control`.
# resource_control:
# # See: https://www.freedesktop.org/software/systemd/man/systemd.resource-control.html#MemoryLimit=bytes
# memory_limit: "2G"
# # See: https://www.freedesktop.org/software/systemd/man/systemd.resource-control.html#CPUQuota=
# # The percentage specifies how much CPU time the unit shall get at maximum, relative to the total CPU time available on one CPU. Use values > 100% for allotting CPU time on more than one CPU.
# # Example: CPUQuota=200% ensures that the executed processes will never get more than two CPU time.
# cpu_quota: "200%"
# # See: https://www.freedesktop.org/software/systemd/man/systemd.resource-control.html#IOReadBandwidthMax=device%20bytes
# io_read_bandwidth_max: "/dev/disk/by-path/pci-0000:00:1f.2-scsi-0:0:0:0 100M"
# io_write_bandwidth_max: "/dev/disk/by-path/pci-0000:00:1f.2-scsi-0:0:0:0 100M"
# # Monitored variables are applied to all the machines.
monitored:
# # The communication port for reporting system information of each node in the TiDB cluster.
node_exporter_port: 9100
# # Blackbox_exporter communication port, used for TiDB cluster port monitoring.
blackbox_exporter_port: 9115
# # Storage directory for deployment files, startup scripts, and configuration files of monitoring components.
# deploy_dir: "/tidb-deploy/monitored-9100"
# # Data storage directory of monitoring components.
# data_dir: "/tidb-data/monitored-9100"
# # Log storage directory of the monitoring component.
# log_dir: "/tidb-deploy/monitored-9100/log"
# # Server configs are used to specify the runtime configuration of TiDB components.
# # All configuration items can be found in TiDB docs:
# # - TiDB: https://docs.pingcap.com/tidb/stable/tidb-configuration-file
# # - TiKV: https://docs.pingcap.com/tidb/stable/tikv-configuration-file
# # - PD: https://docs.pingcap.com/tidb/stable/pd-configuration-file
# # - TiFlash: https://docs.pingcap.com/tidb/stable/tiflash-configuration
# #
# # All configuration items use points to represent the hierarchy, e.g:
# # readpool.storage.use-unified-pool
# # ^ ^
# # - example: https://github.com/pingcap/tiup/blob/master/embed/examples/cluster/topology.example.yaml
# # You can overwrite this configuration via the instance-level `config` field.
# server_configs:
# tidb:
# tikv:
# pd:
# tiflash:
# tiflash-learner:
# kvcdc:
# # Server configs are used to specify the configuration of PD Servers.
pd_servers:
# # The ip address of the PD Server.
- host: 10.0.1.11
# # SSH port of the server.
# ssh_port: 22
# # PD Server name
# name: "pd-1"
# # communication port for TiDB Servers to connect.
# client_port: 2379
# # communication port among PD Server nodes.
# peer_port: 2380
# # PD Server deployment file, startup script, configuration file storage directory.
# deploy_dir: "/tidb-deploy/pd-2379"
# # PD Server data storage directory.
# data_dir: "/tidb-data/pd-2379"
# # PD Server log file storage directory.
# log_dir: "/tidb-deploy/pd-2379/log"
# # numa node bindings.
# numa_node: "0,1"
# # The following configs are used to overwrite the `server_configs.pd` values.
# config:
# schedule.max-merge-region-size: 20
# schedule.max-merge-region-keys: 200000
- host: 10.0.1.12
# ssh_port: 22
# name: "pd-1"
# client_port: 2379
# peer_port: 2380
# deploy_dir: "/tidb-deploy/pd-2379"
# data_dir: "/tidb-data/pd-2379"
# log_dir: "/tidb-deploy/pd-2379/log"
# numa_node: "0,1"
# config:
# schedule.max-merge-region-size: 20
# schedule.max-merge-region-keys: 200000
- host: 10.0.1.13
# ssh_port: 22
# name: "pd-1"
# client_port: 2379
# peer_port: 2380
# deploy_dir: "/tidb-deploy/pd-2379"
# data_dir: "/tidb-data/pd-2379"
# log_dir: "/tidb-deploy/pd-2379/log"
# numa_node: "0,1"
# config:
# schedule.max-merge-region-size: 20
# schedule.max-merge-region-keys: 200000
# # Server configs are used to specify the configuration of TiDB Servers.
tidb_servers:
# # The ip address of the TiDB Server.
- host: 10.0.1.14
# # SSH port of the server.
# ssh_port: 22
# # Access the TiDB cluster port.
port: 4000
# # TiDB Server status information reporting port.
status_port: 10080
# # TiDB Server deployment file, startup script, configuration file storage directory.
deploy_dir: "/tidb-deploy/tidb-4000"
# # TiDB Server log file storage directory.
log_dir: "/tidb-deploy/tidb-4000/log"
# numa_node: "0" # suggest numa node bindings.
- host: 10.0.1.14
# ssh_port: 22
port: 4001
status_port: 10081
deploy_dir: "/tidb-deploy/tidb-4001"
log_dir: "/tidb-deploy/tidb-4001/log"
# numa_node: "1" # suggest numa node bindings.
- host: 10.0.1.15
# ssh_port: 22
port: 4000
status_port: 10080
deploy_dir: "/tidb-deploy/tidb-4000"
log_dir: "/tidb-deploy/tidb-4000/log"
# numa_node: "0" # suggest numa node bindings.
- host: 10.0.1.15
# ssh_port: 22
port: 4001
status_port: 10081
deploy_dir: "/tidb-deploy/tidb-4001"
log_dir: "/tidb-deploy/tidb-4001/log"
# numa_node: "1" # suggest numa node bindings.
# # Server configs are used to specify the configuration of TiKV Servers.
tikv_servers:
# # The ip address of the TiKV Server.
- host: 10.0.1.16
# # SSH port of the server.
# ssh_port: 22
# # TiKV Server communication port.
port: 20160
# # Communication port for reporting TiKV Server status.
status_port: 20180
# # TiKV Server deployment file, startup script, configuration file storage directory.
deploy_dir: "/data1/tidb-deploy/tikv-20160"
# # TiKV Server data storage directory.
data_dir: "/data1/tidb-data/tikv-20160"
# # TiKV Server log file storage directory.
log_dir: "/data1/tidb-deploy/tikv-20160/log"
# numa_node: "0"
# # The following configs are used to overwrite the `server_configs.tikv` values.
# config:
# log.level: warn
- host: 10.0.1.16
# ssh_port: 22
port: 20161
status_port: 20181
deploy_dir: "/data2/tidb-deploy/tikv-20161"
data_dir: "/data2/tidb-data/tikv-20161"
log_dir: "/data2/tidb-deploy/tikv-20161/log"
# numa_node: "1"
# config:
# log.level: warn
- host: 10.0.1.17
# ssh_port: 22
port: 20160
status_port: 20180
deploy_dir: "/data1/tidb-deploy/tikv-20160"
data_dir: "/data1/tidb-data/tikv-20160"
log_dir: "/data1/tidb-deploy/tikv-20160/log"
# numa_node: "0"
# config:
# log.level: warn
- host: 10.0.1.17
# ssh_port: 22
port: 20161
status_port: 20181
deploy_dir: "/data2/tidb-deploy/tikv-20161"
data_dir: "/data2/tidb-data/tikv-20161"
log_dir: "/data2/tidb-deploy/tikv-20161/log"
# numa_node: "1"
# config:
# log.level: warn
- host: 10.0.1.18
# ssh_port: 22
port: 20160
status_port: 20180
deploy_dir: "/data1/tidb-deploy/tikv-20160"
data_dir: "/data1/tidb-data/tikv-20160"
log_dir: "/data1/tidb-deploy/tikv-20160/log"
# numa_node: "0"
# config:
# log.level: warn
- host: 10.0.1.18
# ssh_port: 22
port: 20161
status_port: 20181
deploy_dir: "/data2/tidb-deploy/tikv-20161"
data_dir: "/data2/tidb-data/tikv-20161"
log_dir: "/data2/tidb-deploy/tikv-20161/log"
# numa_node: "1"
# config:
# log.level: warn
# # Server configs are used to specify the configuration of TiFlash Servers.
tiflash_servers:
# # The ip address of the TiFlash Server.
- host: 10.0.1.19
# # SSH port of the server.
# ssh_port: 22
# # TiFlash TCP Service port.
# # Since 7.1.0, it is not actually listened, and only being used as part of the instance identity.
tcp_port: 9000
# # TiFlash raft service and coprocessor service listening address.
flash_service_port: 3930
# # TiFlash Proxy service port.
flash_proxy_port: 20170
# # Prometheus pulls TiFlash Proxy metrics port.
flash_proxy_status_port: 20292
# # Prometheus pulls the TiFlash metrics port.
metrics_port: 8234
# # TiFlash Server deployment file, startup script, configuration file storage directory.
deploy_dir: /data1/tidb-deploy/tiflash-9000
## With cluster version >= v4.0.9 and you want to deploy a multi-disk TiFlash node, it is recommended to
## check config.storage.* for details. The data_dir will be ignored if you defined those configurations.
## Setting data_dir to a ','-joined string is still supported but deprecated.
## Check https://docs.pingcap.com/tidb/stable/tiflash-configuration#multi-disk-deployment for more details.
# # TiFlash Server data storage directory.
data_dir: /data1/tidb-data/tiflash-9000
# # TiFlash Server log file storage directory.
log_dir: /data1/tidb-deploy/tiflash-9000/log
# # The following configs are used to overwrite the `server_configs.tiflash` values.
# config:
# logger.level: info
# # The following configs are used to overwrite the `server_configs.tiflash-learner` values.
# learner_config:
# log.level: info
- host: 10.0.1.19
# ssh_port: 22
tcp_port: 9001
flash_service_port: 3931
flash_proxy_port: 20171
flash_proxy_status_port: 20293
metrics_port: 8235
deploy_dir: /data2/tidb-deploy/tiflash-9001
data_dir: /data2/tidb-data/tiflash-9001
log_dir: /data2/tidb-deploy/tiflash-9001/log
- host: 10.0.1.20
# ssh_port: 22
tcp_port: 9000
flash_service_port: 3930
flash_proxy_port: 20170
flash_proxy_status_port: 20292
metrics_port: 8234
deploy_dir: /data1/tidb-deploy/tiflash-9000
data_dir: /data1/tidb-data/tiflash-9000
log_dir: /data1/tidb-deploy/tiflash-9000/log
- host: 10.0.1.20
# ssh_port: 22
tcp_port: 9001
flash_service_port: 3931
flash_proxy_port: 20171
flash_proxy_status_port: 20293
metrics_port: 8235
deploy_dir: /data2/tidb-deploy/tiflash-9001
data_dir: /data2/tidb-data/tiflash-9001
log_dir: /data2/tidb-deploy/tiflash-9001/log
# # Server configs are used to specify the configuration of TiKV-CDC Servers.
kvcdc_servers:
- host: 10.0.1.20
# # SSH port of the server.
# ssh_port: 22
# # TiKV-CDC Server communication port.
port: 8600
# # TiKV-CDC Server data storage directory.
data_dir: "/data1/tidb-data/tikv-cdc-8600"
# # TiKV-CDC Server log file storage directory.
log_dir: "/data1/tidb-deploy/tikv-cdc-8600/log"
- host: 10.0.1.21
data_dir: "/data1/tidb-data/tikv-cdc-8600"
log_dir: "/data1/tidb-deploy/tikv-cdc-8600/log"
# # Server configs are used to specify the configuration of TiCDC Servers.
#cdc_servers:
# - host: 10.0.1.20
# # SSH port of the server.
# ssh_port: 22
# # TiCDC Server communication port.
# port: 8300
# # TiCDC Server data storage directory.
# data_dir: "/data1/tidb-deploy/cdc-8300"
# # TiCDC Server log file storage directory.
# log_dir: "/data1/tidb-deploy/cdc-8300/log"
# # TiCDC Server deployment file, startup script, configuration file storage directory.
# deploy_dir: "/data1/tidb-deploy/cdc-8300"
# gc-ttl: 86400 # 24h
# tz: "System"
# numa_node: "0,1"
# config:
# log.level: warn
# ticdc_cluster_id: "default"
# - host: 10.0.1.21
# ssh_port: 22
# port: 8301
# data_dir: "/data2/tidb-deploy/cdc-8301"
# log_dir: "/data2/tidb-deploy/cdc-8301/log"
# deploy_dir: "/data2/tidb-deploy/cdc-8301"
# # Server configs are used to specify the configuration of TiDB Dashboard Servers. Available from v6.5.0
# tidb_dashboard_servers:
# # The ip address of the PD Server.
# - host: 10.0.1.11
# # SSH port of the server.
# ssh_port: 22
# # port of TiDB Dashboard
# port: 12333
# # TiDB Dashboard deployment file, startup script, configuration file storage directory.
# deploy_dir: "/tidb-deploy/tidb-dashboard-12333"
# # PD Server data storage directory.
# data_dir: "/tidb-data/tidb-dashboard-12333"
# # PD Server log file storage directory.
# log_dir: "/tidb-deploy/tidb-dashboard-12333/log"
# # numa node bindings.
# numa_node: "0,1"
# # Server configs are used to specify the configuration of Prometheus Server.
monitoring_servers:
# # The ip address of the Monitoring Server.
- host: 10.0.1.21
# # SSH port of the server.
# ssh_port: 22
# # Prometheus Service communication port.
# port: 9090
# # ng-monitoring servive communication port
# ng_port: 12020
# # Prometheus deployment file, startup script, configuration file storage directory.
# deploy_dir: "/tidb-deploy/prometheus-8249"
# # Prometheus data storage directory.
# data_dir: "/tidb-data/prometheus-8249"
# # Prometheus log file storage directory.
# log_dir: "/tidb-deploy/prometheus-8249/log"
# prometheus rule dir on TiUP machine
# rule_dir: /home/tidb/prometheus_rule
# scrape_interval: 15s
# scrape_timeout: 10s
# # Server configs are used to specify the configuration of Grafana Servers.
grafana_servers:
# # The ip address of the Grafana Server.
- host: 10.0.1.21
# # Grafana Web monitoring service client (browser) access port
# port: 3000
# # Grafana deployment file, startup script, configuration file storage directory.
# deploy_dir: /tidb-deploy/grafana-3000
# grafana dashboard dir on TiUP machine
# dashboard_dir: /home/tidb/dashboards
# config:
# log.file.level: warning
# # Server configs are used to specify the configuration of Alertmanager Servers.
alertmanager_servers:
# # The ip address of the Alertmanager Server.
- host: 10.0.1.21
# # SSH port of the server.
# ssh_port: 22
# Alertmanager web service listen host.
# listen_host: 0.0.0.0
# # Alertmanager web service port.
# web_port: 9093
# # Alertmanager communication port.
# cluster_port: 9094
# # Alertmanager deployment file, startup script, configuration file storage directory.
# deploy_dir: "/tidb-deploy/alertmanager-9093"
# # Alertmanager data storage directory.
# data_dir: "/tidb-data/alertmanager-9093"
# # Alertmanager log file storage directory.
# log_dir: "/tidb-deploy/alertmanager-9093/log"
# # Alertmanager config file storage directory.
# config_file: "/tidb-deploy/alertmanager-9093/bin/alertmanager/alertmanager.yml"