This is some example code for how to use Hashicorp's Raft implementation with gRPC.
This example uses Hashicorp's Raft Transport to communicate between nodes using TCP.
This example use two independent ports support application's service and raft's transport.
GRPC Server port: 40051, 40052, 40053, 40054 ... RAFT Server port: 50051, 50052, 50053, 50054 ...
$ mkdir /tmp/test
$ mkdir /tmp/test/node{A,B,C}
$ ./raft-example --raft_bootstrap --raft_id=nodeA --grpc_addr=localhost:40051 --raft_addr=localhost:50051 --raft_data_dir /tmp/test
$ ./raft-example --raft_id=nodeB --grpc_addr=localhost:40052 --raft_addr=localhost:50052 --raft_data_dir /tmp/test
$ ./raft-example --raft_id=nodeC --grpc_addr=localhost:40053 --raft_addr=localhost:50053 --raft_data_dir /tmp/test
You start up three nodes, and bootstrap one of them. Then you tell the bootstrapped node where to find peers. Those peers sync up to the state of the bootstrapped node and become members of the cluster. Once your cluster is running, you never need to pass --raft_bootstrap
again.
raft-manager is used to communicate with the cluster and add the other nodes.
add nodes into your cluster, first call applied_index
, then add_voter
.
$ go install github.com/xkeyideal/raft-manager/cmd/manager@latest
$ ./manager localhost:40051 applied_index
$ ./manager localhost:40051 add_voter nodeB localhost:50052 132
$ ./manager localhost:40051 applied_index
$ ./manager localhost:40051 add_voter nodeC localhost:50053 133
exec cmd.go test the raft write and read.
$ go run cmd/cmd.go
Raft uses logs to synchronize changes. Every change submitted to a Raft cluster is a log entry, which gets stored and replicated to the followers in the cluster. In this example, we use raft-pebbledb to store these logs. Once in a while Raft decides the logs have grown too large, and makes a snapshot. Your code is asked to write out its state. That state captures all previous logs. Now Raft can delete all the old logs and just use the snapshot. These snapshots are stored using the FileSnapshotStore, which means they'll just be files in your disk.
You can see all this happening in NewRaft()
in engine/raft.go
.
See fsm/fsm.go
. You'll need to implement a raft.FSM
, and you probably want a gRPC RPC interface.
The Apply
method of Hashicorp Raft must be called by leader.
// Apply is used to apply a command to the FSM in a highly consistent
// manner. This returns a future that can be used to wait on the application.
// An optional timeout can be provided to limit the amount of time we wait
// for the command to be started. This must be run on the leader or it
// will fail.
func (r *Raft) Apply(cmd []byte, timeout time.Duration) ApplyFuture {
return r.ApplyLog(Log{Data: cmd}, timeout)
}
So we need forward Apply
request to the leader. Like the Consul KV https://github.com/hashicorp/consul/blob/main/agent/consul/kvs_endpoint.go#L100 by RPC.
We must know the leader RPC address not raft address when want to forward request to leader.
We can use gossip protocol propagate per node RPC, raft address infos and so on. Best practices of Gossip by Memberlist
The Raft Example use constant map, you can modify it in https://github.com/xkeyideal/raft-example/blob/master/service/service.go
var (
server_lookup = map[string]string{
"127.0.0.1:50051": "127.0.0.1:40051",
"127.0.0.1:50052": "127.0.0.1:40052",
"127.0.0.1:50053": "127.0.0.1:40053",
}
)
- SnapshotInterval & SnapshotThreshold,
SnapshotInterval
controls how often we check if we should perform a snapshot.SnapshotThreshold
controls how many outstanding logs there must be before we perform a snapshot.// runSnapshots is a long running goroutine used to manage taking // new snapshots of the FSM. It runs in parallel to the FSM and // main goroutines, so that snapshots do not block normal operation. func (r *Raft) runSnapshots() { for { select { case <-randomTimeout(r.config().SnapshotInterval): // Check if we should snapshot if !r.shouldSnapshot() { continue } // Trigger a snapshot if _, err := r.takeSnapshot(); err != nil { r.logger.Error("failed to take snapshot", "error", err) } case future := <-r.userSnapshotCh: // User-triggered, run immediately id, err := r.takeSnapshot() if err != nil { r.logger.Error("failed to take snapshot", "error", err) } else { future.opener = func() (*SnapshotMeta, io.ReadCloser, error) { return r.snapshots.Open(id) } } future.respond(err) case <-r.shutdownCh: return } } }
- HeartbeatTimeout, the time in follower state without contact from a leader before we attempt an election.
- ElectionTimeout, the time in candidate state without contact from a leader before we attempt an election.
- TrailingLogs, controls how many logs we leave after a snapshot. This is used so that we can quickly replay logs on a follower instead of being forced to send an entire snapshot.
So the most critical parameters of raft for followers sync logs are SnapshotThreshold and TrailingLogs, and the value of SnapshotThreshold should be less than the value of TrailingLogs.
- NoSnapshotRestoreOnStart, the default to
false
, will restoreSnapshot whenNewRaft
.
rqlite is an easy-to-use, lightweight, distributed relational database, which uses SQLite as its storage engine.
raft-example is under the BSD 2-Clause License. See the LICENSE file for details.