Welcome to mirror list, hosted at ThFree Co, Russian Federation.

gitlab.com/gitlab-org/gitlab-foss.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'doc/administration/consul.md')
-rw-r--r--doc/administration/consul.md12
1 files changed, 6 insertions, 6 deletions
diff --git a/doc/administration/consul.md b/doc/administration/consul.md
index e1f2bd90e05..e282960c857 100644
--- a/doc/administration/consul.md
+++ b/doc/administration/consul.md
@@ -80,14 +80,14 @@ Nodes should be:
- Upgraded one node at a time.
Identify any existing health issues in the cluster by running the following command
-within each node. The command will return an empty array if the cluster is healthy:
+in each node. The command returns an empty array if the cluster is healthy:
```shell
curl "http://127.0.0.1:8500/v1/health/state/critical"
```
-If the Consul version has changed, you'll see a notice at the end of `gitlab-ctl reconfigure`
-informing you that Consul needs to be restarted for the new version to be used.
+If the Consul version has changed, you see a notice at the end of `gitlab-ctl reconfigure`
+informing you that Consul must be restarted for the new version to be used.
Restart Consul one node at a time:
@@ -96,7 +96,7 @@ sudo gitlab-ctl restart consul
```
Consul nodes communicate using the raft protocol. If the current leader goes
-offline, there needs to be a leader election. A leader node must exist to facilitate
+offline, there must be a leader election. A leader node must exist to facilitate
synchronization across the cluster. If too many nodes go offline at the same time,
the cluster loses quorum and doesn't elect a leader due to
[broken consensus](https://www.consul.io/docs/architecture/consensus).
@@ -111,7 +111,7 @@ bundled Consul wasn't used by any process other than GitLab itself, you can
## Troubleshooting Consul
-Below are some useful operations should you need to debug any issues.
+Below are some operations should you debug any issues.
You can see any error logs by running:
```shell
@@ -149,7 +149,7 @@ To be safe, it's recommended that you only restart Consul in one node at a time
ensure the cluster remains intact. For larger clusters, it is possible to restart
multiple nodes at a time. See the
[Consul consensus document](https://www.consul.io/docs/architecture/consensus#deployment-table)
-for the number of failures it can tolerate. This will be the number of simultaneous
+for the number of failures it can tolerate. This is the number of simultaneous
restarts it can sustain.
To restart Consul: