These examples will use a single machine cluster to show you the basics of the etcd REST API. Let's start etcd:
./bin/etcd -data-dir machine0 -name machine0
This will bring up etcd listening on default ports (4001 for client communication and 7001 for server-to-server communication).
The -data-dir machine0
argument tells etcd to write machine configuration, logs and snapshots to the ./machine0/
directory.
The -name machine0
tells the rest of the cluster that this machine is named machine0.
The etcd version of a specific instance can be obtained from the /version
endpoint.
curl -L http://127.0.0.1:4001/version
The primary API of etcd is a hierarchical key space. The key space consists of directories and keys which are generically referred to as "nodes".
Let's set the first key-value pair in the datastore.
In this case the key is /message
and the value is Hello world
.
curl -L http://127.0.0.1:4001/v2/keys/message -XPUT -d value="Hello world"
{
"action": "set",
"node": {
"createdIndex": 2,
"key": "/message",
"modifiedIndex": 2,
"value": "Hello world"
}
}
The response object contains several attributes:
action
: the action of the request that was just made.
The request attempted to modify node.value
via a PUT
HTTP request, thus the value of action is set
.
node.key
: the HTTP path to which the request was made.
We set /message
to Hello world
, so the key field is /message
.
etcd uses a file-system-like structure to represent the key-value pairs, therefore all keys start with /
.
node.value
: the value of the key after resolving the request.
In this case, a successful request was made that attempted to change the node's value to Hello world
.
node.createdIndex
: an index is a unique, monotonically-incrementing integer created for each change to etcd.
This specific index reflects the point in the etcd state machine at which a given key was created.
You may notice that in this example the index is 2
even though it is the first request you sent to the server.
This is because there are internal commands that also change the state behind the scenes, like adding and syncing servers.
node.modifiedIndex
: like node.createdIndex
, this attribute is also an etcd index.
Actions that cause the value to change include set
, delete
, update
, create
, compareAndSwap
and compareAndDelete
.
Since the get
and watch
commands do not change state in the store, they do not change the value of node.modifiedIndex
.
etcd includes a few HTTP headers in responses that provide global information about the etcd cluster that serviced a request:
X-Etcd-Index: 35
X-Raft-Index: 5398
X-Raft-Term: 0
X-Etcd-Index
is the current etcd index as explained above.X-Raft-Index
is similar to the etcd index but is for the underlying raft protocolX-Raft-Term
is an integer that will increase whenever an etcd master election happens in the cluster. If this number is increasing rapidly, you may need to tune the election timeout. See the tuning section for details.We can get the value that we just set in /message
by issuing a GET
request:
curl -L http://127.0.0.1:4001/v2/keys/message
{
"action": "get",
"node": {
"createdIndex": 2,
"key": "/message",
"modifiedIndex": 2,
"value": "Hello world"
}
}
You can change the value of /message
from Hello world
to Hello etcd
with another PUT
request to the key:
curl -L http://127.0.0.1:4001/v2/keys/message -XPUT -d value="Hello etcd"
{
"action": "set",
"node": {
"createdIndex": 3,
"key": "/message",
"modifiedIndex": 3,
"value": "Hello etcd"
},
"prevNode": {
"createdIndex": 2
"key": "/message",
"value": "Hello world",
"modifiedIndex": 2,
}
}
Here we introduce a new field: prevNode
. The prevNode
field represents what the state of a given node was before resolving the request at hand. The prevNode
field follows the same format as the node
, and is omitted in the event that there was no previous state for a given node.
You can remove the /message
key with a DELETE
request:
curl -L http://127.0.0.1:4001/v2/keys/message -XDELETE
{
"action": "delete",
"node": {
"createdIndex": 3,
"key": "/message",
"modifiedIndex": 4
},
"prevNode": {
"key": "/message",
"value": "Hello etcd",
"modifiedIndex": 3,
"createdIndex": 3
}
}
Keys in etcd can be set to expire after a specified number of seconds.
You can do this by setting a TTL (time to live) on the key when sending a PUT
request:
curl -L http://127.0.0.1:4001/v2/keys/foo -XPUT -d value=bar -d ttl=5
{
"action": "set",
"node": {
"createdIndex": 5,
"expiration": "2013-12-04T12:01:21.874888581-08:00",
"key": "/foo",
"modifiedIndex": 5,
"ttl": 5,
"value": "bar"
}
}
Note the two new fields in response:
The expiration
is the time at which this key will expire and be deleted.
The ttl
is the specified time to live for the key, in seconds.
NOTE: Keys can only be expired by a cluster leader, so if a machine gets disconnected from the cluster, its keys will not expire until it rejoins.
Now you can try to get the key by sending a GET
request:
curl -L http://127.0.0.1:4001/v2/keys/foo
If the TTL has expired, the key will have been deleted, and you will be returned a 100.
{
"cause": "/foo",
"errorCode": 100,
"index": 6,
"message": "Key Not Found"
}
The TTL could be unset to avoid expiration through update operation:
curl -L http://127.0.0.1:4001/v2/keys/foo -XPUT -d value=bar -d ttl= -d prevExist=true
{
"action": "update",
"node": {
"createdIndex": 5,
"key": "/foo",
"modifiedIndex": 6,
"value": "bar"
}
"prevNode": {
"createdIndex": 5,
"expiration": "2013-12-04T12:01:21.874888581-08:00",
"key": "/foo",
"modifiedIndex": 5,
"ttl": 3,
"value": "bar"
}
}
We can watch for a change on a key and receive a notification by using long polling.
This also works for child keys by passing recursive=true
in curl.
In one terminal, we send a GET
with wait=true
:
curl -L http://127.0.0.1:4001/v2/keys/foo?wait=true
Now we are waiting for any changes at path /foo
.
In another terminal, we set a key /foo
with value bar
:
curl -L http://127.0.0.1:4001/v2/keys/foo -XPUT -d value=bar
The first terminal should get the notification and return with the same response as the set request:
{
"action": "set",
"node": {
"createdIndex": 7,
"key": "/foo",
"modifiedIndex": 7,
"value": "bar"
},
"prevNode": {
"createdIndex": 6,
"key": "/foo",
"modifiedIndex": 6,
"value": "bar"
}
}
However, the watch command can do more than this. Using the index, we can watch for commands that have happened in the past. This is useful for ensuring you don't miss events between watch commands.
Let's try to watch for the set command of index 7 again:
curl -L 'http://127.0.0.1:4001/v2/keys/foo?wait=true&waitIndex=7'
The watch command returns immediately with the same response as previously.
Using POST
on a directory, you can create keys with key names that are created in-order.
This can be used in a variety of useful patterns, like implementing queues of keys which need to be processed in strict order.
An example use case is the locking module which uses it to ensure clients get fair access to a mutex.
Creating an in-order key is easy:
curl http://127.0.0.1:4001/v2/keys/queue -XPOST -d value=Job1
{
"action": "create",
"node": {
"createdIndex": 6,
"key": "/queue/6",
"modifiedIndex": 6,
"value": "Job1"
}
}
If you create another entry some time later, it is guaranteed to have a key name that is greater than the previous key.
Also note the key names use the global etcd index, so the next key can be more than previous + 1
.
curl http://127.0.0.1:4001/v2/keys/queue -XPOST -d value=Job2
{
"action": "create",
"node": {
"createdIndex": 29,
"key": "/queue/29",
"modifiedIndex": 29,
"value": "Job2"
}
}
To enumerate the in-order keys as a sorted list, use the "sorted" parameter.
curl -s 'http://127.0.0.1:4001/v2/keys/queue?recursive=true&sorted=true'
{
"action": "get",
"node": {
"createdIndex": 2,
"dir": true,
"key": "/queue",
"modifiedIndex": 2,
"nodes": [
{
"createdIndex": 2,
"key": "/queue/2",
"modifiedIndex": 2,
"value": "Job1"
},
{
"createdIndex": 3,
"key": "/queue/3",
"modifiedIndex": 3,
"value": "Job2"
}
]
}
}
Like keys, directories in etcd can be set to expire after a specified number of seconds.
You can do this by setting a TTL (time to live) on a directory when it is created with a PUT
:
curl -L http://127.0.0.1:4001/v2/keys/dir -XPUT -d ttl=30 -d dir=true
{
"action": "set",
"node": {
"createdIndex": 17,
"dir": true,
"expiration": "2013-12-11T10:37:33.689275857-08:00",
"key": "/newdir",
"modifiedIndex": 17,
"ttl": 30
}
}
The directory's TTL can be refreshed by making an update.
You can do this by making a PUT with prevExist=true
and a new TTL.
curl -L http://127.0.0.1:4001/v2/keys/dir -XPUT -d ttl=30 -d dir=true -d prevExist=true
Keys that are under this directory work as usual, but when the directory expires, a watcher on a key under the directory will get an expire event:
curl 'http://127.0.0.1:4001/v2/keys/dir/asdf?consistent=true&wait=true'
{
"action": "expire",
"node": {
"createdIndex": 8,
"key": "/dir",
"modifiedIndex": 15
},
"prevNode": {
"createdIndex": 8,
"key": "/dir",
"dir":true,
"modifiedIndex": 17,
"expiration": "2013-12-11T10:39:35.689275857-08:00"
},
}
etcd can be used as a centralized coordination service in a cluster, and CompareAndSwap
(CAS) is the most basic operation used to build a distributed lock service.
This command will set the value of a key only if the client-provided conditions are equal to the current conditions.
The current comparable conditions are:
prevValue
- checks the previous value of the key.
prevIndex
- checks the previous modifiedIndex of the key.
prevExist
- checks existence of the key: if prevExist
is true, it is an update
request; if prevExist is false
, it is a create
request.
Here is a simple example.
Let's create a key-value pair first: foo=one
.
curl -L http://127.0.0.1:4001/v2/keys/foo -XPUT -d value=one
Now let's try some invalid CompareAndSwap
commands.
Trying to set this existing key with prevExist=false
fails as expected:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevExist=false -XPUT -d value=three
The error code explains the problem:
{
"cause": "/foo",
"errorCode": 105,
"index": 39776,
"message": "Already exists"
}
Now let's provide a prevValue
parameter:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevValue=two -XPUT -d value=three
This will try to compare the previous value of the key and the previous value we provided. If they are equal, the value of the key will change to three.
{
"cause": "[two != one]",
"errorCode": 101,
"index": 8,
"message": "Test Failed"
}
which means CompareAndSwap
failed. cause
explains why the test failed.
Note: the condition prevIndex=0 always passes.
Let's try a valid condition:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevValue=one -XPUT -d value=two
The response should be:
{
"action": "compareAndSwap",
"node": {
"createdIndex": 8,
"key": "/foo",
"modifiedIndex": 9,
"value": "two"
},
"prevNode": {
"createdIndex": 8,
"key": "/foo",
"modifiedIndex": 8,
"value": "one"
}
}
We successfully changed the value from "one" to "two" since we gave the correct previous value.
This command will delete a key only if the client-provided conditions are equal to the current conditions.
The current comparable conditions are:
prevValue
- checks the previous value of the key.
prevIndex
- checks the previous modifiedIndex of the key.
Here is a simple example. Let's first create a key: foo=one
.
curl -L http://127.0.0.1:4001/v2/keys/foo -XPUT -d value=one
Now let's try some CompareAndDelete
commands.
Trying to delete the key with prevValue=two
fails as expected:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevValue=two -XDELETE
The error code explains the problem:
{
"errorCode": 101,
"message": "Compare failed",
"cause": "[two != one]",
"index": 8
}
As does a CompareAndDelete
with a mismatched prevIndex
:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevIndex=1 -XDELETE
{
"errorCode": 101,
"message": "Compare failed",
"cause": "[1 != 8]",
"index": 8
}
And now a valid prevValue
condition:
curl -L http://127.0.0.1:4001/v2/keys/foo?prevValue=one -XDELETE
The successful response will look something like:
{
"action": "compareAndDelete",
"node": {
"key": "/foo",
"modifiedIndex": 9,
"createdIndex": 8
},
"prevNode": {
"key": "/foo",
"value": "one",
"modifiedIndex": 8,
"createdIndex": 8
}
}
In most cases, directories for a key are automatically created. But there are cases where you will want to create a directory or remove one.
Creating a directory is just like a key except you cannot provide a value and must add the dir=true
parameter.
curl -L http://127.0.0.1:4001/v2/keys/dir -XPUT -d dir=true
{
"action": "set",
"node": {
"createdIndex": 30,
"dir": true,
"key": "/dir",
"modifiedIndex": 30
}
}
In etcd we can store two types of things: keys and directories. Keys store a single string value. Directories store a set of keys and/or other directories.
In this example, let's first create some keys:
We already have /foo=two
so now we'll create another one called /foo_dir/foo
with the value of bar
:
curl -L http://127.0.0.1:4001/v2/keys/foo_dir/foo -XPUT -d value=bar
{
"action": "set",
"node": {
"createdIndex": 2,
"key": "/foo_dir/foo",
"modifiedIndex": 2,
"value": "bar"
}
}
Now we can list the keys under root /
:
curl -L http://127.0.0.1:4001/v2/keys/
We should see the response as an array of items:
{
"action": "get",
"node": {
"dir": true,
"key": "/",
"nodes": [
{
"createdIndex": 2,
"dir": true,
"key": "/foo_dir",
"modifiedIndex": 2
}
]
}
}
Here we can see /foo
is a key-value pair under /
and /foo_dir
is a directory.
We can also recursively get all the contents under a directory by adding recursive=true
.
curl -L http://127.0.0.1:4001/v2/keys/?recursive=true
{
"action": "get",
"node": {
"dir": true,
"key": "/",
"nodes": [
{
"createdIndex": 2,
"dir": true,
"key": "/foo_dir",
"modifiedIndex": 2,
"nodes": [
{
"createdIndex": 2,
"key": "/foo_dir/foo",
"modifiedIndex": 2,
"value": "bar"
}
]
}
]
}
}
Now let's try to delete the directory /foo_dir
.
You can remove an empty directory using the DELETE
verb and the dir=true
parameter.
curl -L 'http://127.0.0.1:4001/v2/keys/foo_dir?dir=true' -XDELETE
{
"action": "delete",
"node": {
"createdIndex": 30,
"dir": true,
"key": "/foo_dir",
"modifiedIndex": 31
},
"prevNode": {
"createdIndex": 30,
"key": "/foo_dir",
"dir": true,
"modifiedIndex": 30
}
}
To delete a directory that holds keys, you must add recursive=true
.
curl -L http://127.0.0.1:4001/v2/keys/dir?recursive=true -XDELETE
{
"action": "delete",
"node": {
"createdIndex": 10,
"dir": true,
"key": "/dir",
"modifiedIndex": 11
},
"prevNode": {
"createdIndex": 10,
"dir": true,
"key": "/dir",
"modifiedIndex": 10
}
}
We can create a hidden key-value pair or directory by add a _
prefix.
The hidden item will not be listed when sending a GET
request for a directory.
First we'll add a hidden key named /_message
:
curl -L http://127.0.0.1:4001/v2/keys/_message -XPUT -d value="Hello hidden world"
{
"action": "set",
"node": {
"createdIndex": 3,
"key": "/_message",
"modifiedIndex": 3,
"value": "Hello hidden world"
}
}
Next we'll add a regular key named /message
:
curl -L http://127.0.0.1:4001/v2/keys/message -XPUT -d value="Hello world"
{
"action": "set",
"node": {
"createdIndex": 4,
"key": "/message",
"modifiedIndex": 4,
"value": "Hello world"
}
}
Now let's try to get a listing of keys under the root directory, /
:
curl -L http://127.0.0.1:4001/v2/keys/
{
"action": "get",
"node": {
"dir": true,
"key": "/",
"nodes": [
{
"createdIndex": 2,
"dir": true,
"key": "/foo_dir",
"modifiedIndex": 2
},
{
"createdIndex": 4,
"key": "/message",
"modifiedIndex": 4,
"value": "Hello world"
}
]
}
}
Here we see the /message
key but our hidden /_message
key is not returned.
You can also use etcd to store small configuration files, json documents, XML documents, etc directly. For example you can use curl to upload a simple text file and encode it:
echo "Hello\nWorld" > afile.txt
curl -L http://127.0.0.1:4001/v2/keys/afile -XPUT --data-urlencode value@afile.txt
{
"action": "get",
"node": {
"createdIndex": 2,
"key": "/afile",
"modifiedIndex": 2,
"value": "Hello\nWorld\n"
}
}
Followers in a cluster can be behind the leader in their copy of the keyspace.
If your application wants or needs the most up-to-date version of a key then it should ensure it reads from the current leader.
By using the consistent=true
flag in your GET requests, etcd will make sure you are talking to the current master.
As an example of how a machine can be behind the leader let's start with a three machine cluster: L, F1, and F2. A client makes a write to L and F1 acknowledges the request. The client is told the write was successful and the keyspace is updated. Meanwhile F2 has partitioned from the network and will have an out-of-date version of the keyspace until the partition resolves. Since F2 missed the most recent write, a client reading from F2 will have an out-of-date version of the keyspace.
Implementation notes on consistent=true
: If the leader you are talking to is
partitioned it will be unable to determine if it is not currently the master.
In a later version we will provide a mechanism to set an upperbound of time
that the current master can be unable to contact the quorom and still serve
reads.
If you want a read that is fully linearized you can use a quorum=true
GET.
The read will take a very similar path to a write and will have a similar
speed. If you are unsure if you need this feature feel free to email etcd-dev
for advice.
The lock module is used to serialize access to resources used by clients. Multiple clients can attempt to acquire a lock but only one can have it at a time. Once the lock is released, the next client waiting for the lock will receive it.
Warning: This module is deprecated and removed at v0.4. See Modules for more details.
To acquire a lock, simply send a POST
request to the lock module with the lock name and TTL:
curl -L http://127.0.0.1:4001/mod/v2/lock/mylock -XPOST -d ttl=20
You will receive the lock index when you acquire the lock:
2
If the TTL is not specified or is not a number then you'll receive the following error:
{
"errorCode": 202,
"message": "The given TTL in POST form is not a number",
"cause": "Acquire",
}
If you specify a timeout that is not a number then you'll receive the following error:
{
"errorCode": 205,
"message": "The given timeout in POST form is not a number",
"cause": "Acquire",
}
To extend the TTL of an already acquired lock, simply repeat your original request but with a PUT
and the lock index instead:
curl -L http://127.0.0.1:4001/mod/v2/lock/mylock -XPUT -d index=5 -d ttl=20
If the index or value is not specified then you'll receive the following error:
{
"errorCode": 207,
"message": "Index or value is required",
"cause": "Renew",
}
If the index or value does not exist then you'll receive the following error with a 404 Not Found
HTTP code:
{
"errorCode": 100,
"message": "Key not found",
"index": 1
}
If the TTL is not specified or is not a number then you'll receive the following error:
{
"errorCode": 202,
"message": "The given TTL in POST form is not a number",
"cause": "Renew",
}
When the client is finished with the lock, simply send a DELETE
request to release the lock:
curl -L http://127.0.0.1:4001/mod/v2/lock/mylock?index=5 -XDELETE
If the index or value is not specified then you'll receive the following error:
{
"errorCode": 207,
"message": "Index or value is required",
"cause": "Release",
}
If the index and value are both specified then you'll receive the following error:
{
"errorCode": 208,
"message": "Index and value cannot both be specified",
"cause": "Release",
}
If the index or value does not exist then you'll receive the following error with a 404 Not Found
HTTP code:
{
"errorCode": 100,
"message": "Key not found",
"index": 1
}
To determine the current value or index of a lock, send a GET
request to the lock.
You can specify a field
of index
or value
.
The default is value
.
curl -L http://127.0.0.1:4001/mod/v2/lock/mylock?field=index
Will return the current index:
2
If you specify a field other than index
or value
then you'll receive the following error:
{
"errorCode": 209,
"message": "Invalid field",
"cause": "Get",
}
The leader module wraps the lock module to provide a simple interface for electing a single leader in a cluster.
Warning: This module is deprecated at v0.4. See Modules for more details.
A client can attempt to become leader by sending a PUT
request to the leader module with the name of the leader to elect:
curl -L http://127.0.0.1:4001/mod/v2/leader/myclustername -XPUT -d ttl=300 -d name=foo.mydomain.com
You will receive a successful 200
HTTP response code when the leader is elected.
If the name is not specified then you'll receive the following error:
{
"errorCode": 206,
"message": "Name is required in POST form",
"cause": "Set",
}
You can also receive any errors specified by the Lock module.
A client can check to determine if there is a current leader by sending a GET
request to the leader module:
curl -L http://127.0.0.1:4001/mod/v2/leader/myclustername
You will receive the name of the current leader:
foo.mydomain.com
A client can give up leadership by sending a DELETE
request with the leader name:
curl -L http://127.0.0.1:4001/mod/v2/leader/myclustername?name=foo.mydomain.com -XDELETE
If the name is not specified then you'll receive the following error:
{
"errorCode": 206,
"message": "Name is required in POST form",
"cause": "Set",
}
An etcd cluster keeps track of a number of statistics including latency, bandwidth and uptime.
These statistics are used in the /mod/dashboard
endpoint to generate tables and graphs about the cluster state.
The leader has a view of the entire cluster and keeps track of two interesting statistics: latency to each peer in the cluster, and the number of failed and successful Raft RPC requests.
You can grab these statistics from the /v2/stats/leader
endpoint:
curl -L http://127.0.0.1:4001/v2/stats/leader
{
"followers": {
"etcd-node1": {
"counts": {
"fail": 1212,
"success": 4163176
},
"latency": {
"average": 2.7206299430775007,
"current": 1.486487,
"maximum": 2018.410279,
"minimum": 1.011763,
"standardDeviation": 6.246990702203536
}
},
"etcd-node3": {
"counts": {
"fail": 1378,
"success": 4164598
},
"latency": {
"average": 2.707100125761001,
"current": 1.666258,
"maximum": 1409.054765,
"minimum": 0.998415,
"standardDeviation": 5.910089773061448
}
}
},
"leader": "etcd-node2"
}
Each node keeps a number of internal statistics:
leaderInfo.leader
: name of the current leader machineleaderInfo.uptime
: amount of time the leader has been leadername
: this machine's namerecvAppendRequestCnt
: number of append requests this node has processedrecvBandwidthRate
: number of bytes per second this node is receiving (follower only)recvPkgRate
: number of requests per second this node is receiving (follower only)sendAppendRequestCnt
: number of requests that this node has sentsendBandwidthRate
: number of bytes per second this node is receiving (leader only). This value is undefined on single machine clusters.sendPkgRate
: number of requests per second this node is receiving (leader only). This value is undefined on single machine clusters.state
: either leader or followerstartTime
: the time when this node was startedThis is an example response from a follower machine:
curl -L http://127.0.0.1:4001/v2/stats/self
{
"leaderInfo": {
"leader": "machine1",
"uptime": "1m18.544996775s"
},
"name": "machine0",
"recvAppendRequestCnt": 5871307,
"recvBandwidthRate": 630.3121596542599,
"recvPkgRate": 19.272654323628185,
"sendAppendRequestCnt": 3175763,
"startTime": "2014-01-01T15:26:24.96569404Z",
"state": "follower"
}
And this is an example response from a leader machine:
curl -L http://127.0.0.1:4001/v2/stats/self
{
"leaderInfo": {
"leader": "machine0",
"uptime": "24.648619798s"
},
"name": "machine0",
"recvAppendRequestCnt": 5901116,
"sendAppendRequestCnt": 3212344,
"sendBandwidthRate": 1254.3151237301615,
"sendPkgRate": 38.71342974475808,
"startTime": "2014-01-01T15:26:24.96569404Z",
"state": "leader"
}
The store statistics include information about the operations that this node has handled.
Operations that modify the store's state like create, delete, set and update are seen by the entire cluster and the number will increase on all nodes. Operations like get and watch are node local and will only be seen on this node.
curl -L http://127.0.0.1:4001/v2/stats/store
{
"compareAndSwapFail": 0,
"compareAndSwapSuccess": 0,
"createFail": 0,
"createSuccess": 2,
"deleteFail": 0,
"deleteSuccess": 0,
"expireCount": 0,
"getsFail": 4,
"getsSuccess": 75,
"setsFail": 2,
"setsSuccess": 4,
"updateFail": 0,
"updateSuccess": 0,
"watchers": 0
}
The configuration endpoint manages shared cluster wide properties.
curl -L http://127.0.0.1:7001/v2/admin/config -XPUT -d '{"activeSize":3, "removeDelay":1800,"syncInterval":5}'
{
"activeSize": 3,
"removeDelay": 1800,
"syncInterval":5
}
activeSize
is the maximum number of peers that can join the cluster and participate in the consensus protocol.
The size of cluster is controlled to be around a certain number. If it is not, standby-mode instances will join or peer-mode instances will be removed to make it happen.
removeDelay
indicates the minimum time that a machine has been observed to be unresponsive before it is removed from the cluster.
curl -L http://127.0.0.1:7001/v2/admin/config
{
"activeSize": 3,
"removeDelay": 1800,
"syncInterval":5
}
At times you may want to manually remove a machine. Using the machines endpoint you can find and remove machines.
First, list all the machines in the cluster.
curl -L http://127.0.0.1:7001/v2/admin/machines
[
{
"clientURL": "http://127.0.0.1:4001",
"name": "peer1",
"peerURL": "http://127.0.0.1:7001",
"state": "leader"
},
{
"clientURL": "http://127.0.0.1:4002",
"name": "peer2",
"peerURL": "http://127.0.0.1:7002",
"state": "follower"
},
{
"clientURL": "http://127.0.0.1:4003",
"name": "peer3",
"peerURL": "http://127.0.0.1:7003",
"state": "follower"
}
]
Then take a closer look at the machine you want to remove.
curl -L http://127.0.0.1:7001/v2/admin/machines/peer2
{
"clientURL": "http://127.0.0.1:4002",
"name": "peer2",
"peerURL": "http://127.0.0.1:7002",
"state": "follower"
}
And finally remove it.
curl -L -XDELETE http://127.0.0.1:7001/v2/admin/machines/peer2