Why Do We Use External Persistent Storage for Redis Mesos Containers?
Since Redis is an in-memory database, an instance/service restart will result in loss of data. To counter this, it is always advisable to snapshot the Redis in-memory database from time to time.
This helps Redis instance to recover from the point in time failure.
In DCOS, Redis is deployed as a stateless service. To make it a stateful and persistent data, we can configure local volumes or external volumes.
The disadvantage of having a local volume mapped to Mesos containers is when a slave node goes down, your local volume becomes unavailable, and the data loss occurs.
However, with external persistent volumes, as they are available on each node of the DCOS cluster, a slave node failure does not impact the data availability.
Rex-Ray
REX-Ray is an open source, storage management solution designed to support container runtimes such as Docker and Mesos.
REX-Ray enables stateful applications such as databases to persist and maintain its data after the life cycle of the container has ended. Built-in high availability enables orchestrators such as Docker Swarm, Kubernetes, and Mesos Frameworks like Marathon to automatically orchestrate storage tasks between hosts in a cluster.
Built on top of the libStorage framework, REX-Ray’s simplified architecture consists of a single binary and runs as a stateless service on every host using a configuration file to orchestrate multiple storage platforms.
Objective: To create a Redis service in DC/OS environment with persistent storage.
Warning: The Persistent Volume feature is still in beta Phase for DC/OS Version 1.11.
Prerequisites:
Make sure the rexray service is running and is in a healthy state for the cluster.
Steps:
Click on the Add button in Services component of DC/OS GUI.
Click on JSON Configuration.
Note: For persistent storage, below code should be added in the normal Redis service configuration JSON file to mount external persistent volumes.
Make sure the service is up and in a running state:
If you look closely, the service was suspended and respawned on a different slave node. We populated the database with dummy data and saved the snapshot in the data directory.
When the service did come upon a different node 10.0.3.204, the data persisted and the volume was visible on the new node.
To connect with Redis service, use below host:port in your applications:
redis.marathon.l4lb.thisdcos.directory:6379
Conclusion
We learned about Standalone Redis Service deployment from DCOS catalog on DCOS. Also, we saw how to add Persistent storage to it using RexRay. We also learned how RexRay automatically manages volumes over AWS ebs and how to integrate them in DCOS apps/services. Finally, we saw how other applications can communicate with this Redis service.
In the first part of this blog, we saw how to install standalone redis service on DC/OS with Persistent storage using RexRay and AWS EBS volumes.
A single server is a single point of failure in every system, so to ensure high availability of redis database, we can deploy a master-slave cluster of Redis servers. In this blog, we will see how to setup such 6 node (3 master, 3 slave) Redis cluster and persist data using RexRay and AWS EBS volumes. After that we will see how to import existing data into this cluster.
Redis Cluster
It is form of replicated Redis servers in multi-master architecture. All the data is sharded into 16384 buckets, where every master node is assigned subset of buckets out of them (generally evenly sharded) and each master replicated by its slaves. It provides more resilience and scaling for production grade deployments where heavy workload is expected. Applications can connect to any node in cluster mode and the request will be redirected to respective master node.
Source: Octo
Objective: To create a Redis cluster with number of services in DCOC environment with persistent storage and import the existing Redis dump.rdb data to the cluster.
Prerequisites:
Make sure rexray component is running and is in a healthy state for DCOS cluster.
Steps:
As per Redis doc, the minimal cluster should have at least 3 master and 3 slave nodes, so making it a total 6 Redis services.
All services will use similar json configuration except changes in names of service, external volume, and port mappings.
We will deploy one Redis service for each Redis cluster node and once all services are running, we will form cluster among them.
We will use host network for Redis node containers, for that we will restrict Redis nodes to run on particular node. This will help us to troubleshoot cluster (fixed IP, so we can restart Redis node any time without data loss).
Using host network adds a prerequisites that number of dcos nodes >= number of Redis nodes.
Click on the Add button in Services tab of DCOS UI
Click on JSON configuration
Add below json config for Redis service, change the values which are written in BLOCK letters with # as prefix and suffix.
#NODENAME# – Name of Redis node (Ex. redis-node-1)
#NODEHOSTIP# – IP of dcos node on which this Redis node will run. This ip must be unique for each Redis node. (Ex. 10.2.12.23)
#VOLUMENAME# – Name of persistent volume, Give name to identify volume on AWS EBS (Ex. <dcos cluster=”” name=””>-redis-node-<node number=””>)</node></dcos>
#NODEVIP# – VIP For the Redis node. It must be ‘Redis’ for first Redis node, for others it can be the same as NODENAME (Ex. redis-node-2)
After updating the highlighted fields, copy above json to json configuration box, click on ‘Review & Run’ button in the right corner, this will start the service with above configuration.
Once above service is UP and Running, then repeat the step 2 to 4 for each Redis node with respective values for highlighted fields.
So if we go with 6 node cluster, at the end we will have 6 Redis nodes UP and Running, like:
Note: Since we are using external volume for persistent storage, we can not scale our services, i.e. each service will only one instance max. If we try to scale, we will get below error :
2. Form the Redis cluster between Redis node services:
To create or manage Redis-cluster, first deploy redis-cluster-util container on DCOS using below json config:
If all OK, it will show OK with status, else it will show ERR with the error message.
3. Import existing dump.rdb to Redis cluster
At this point, all the Redis nodes should be empty and each one should have an ID and some assigned slots:
Before reuse an existing dump data, we have to reshard all slots to one instance. We specify the number of slots to move (all, so 16384), the id we move to (here Node 1 – 10.0.1.90:6379) and where we take these slots from (all other nodes).
redis-trib.rb reshard 10.0.1.90:6379
Parameters:
host:port of any node from the cluster.
Output:
It will prompt for number of slots to move – here all. i.e 16384
Receiving node id – here id of node 10.0.1.90:6379 (redis-node-1)
Source node IDs – here all, as we want to shard all slots to one node.
And prompt to proceed – press ‘yes’
Now check again node 10.0.1.90:6379
redis-trib.rb check 10.0.1.90:6379
Parameters: host:port of any node from the cluster.
Output: it will show all (16384) slots moved to node 10.0.1.90:6379
Next step is Importing our existing Redis dump data.
Now copy the existing dump.rdb to our redis-cluster-util container using below steps:
– Copy existing dump.rdb to dcos node on which redis-cluster-util container is running. Can use scp from any other public server to dcos node.
– Now we have dump .rdb in our dcos node, copy this dump.rdb to redis-cluster-util container using below command:
Now we have dump.rdb in our redis-cluster-util container, we can import it to our Redis cluster. Execute and go to the redis-cluster-util container using:
All data transferred. Waiting for the last reply...Last reply received from server.errors: 0, replies: 4341259
as well as this in the Redis server logs:
95086:M01 Mar 21:53:42.071*10000 changes in60 seconds. Saving...95086:M01 Mar 21:53:42.072* Background saving started by pid 9822398223:C01 Mar 21:53:44.277*DB saved on disk
WARNING: Like our Oracle DB instance can have multiple databases, similarly Redis saves keys in keyspaces. Now when Redis is in cluster mode, it does not accept the dumps which has more than one keyspaces. As per documentation: “Redis Cluster does not support multiple databases like the stand alone version of Redis. There is just database 0 and the SELECT command is not allowed. “
So while importing such multi-keyspace Redis dump, server fails while starting on below issue :
23049:M16 Mar 17:21:17.772*DB loaded from disk: 5.222 seconds23049:M16 Mar 17:21:17.772 # You can't have keys in a DB different than DB 0 when in Cluster mode. Exiting.Solution /WA :
There is redis-cli command “MOVE” to move keys from one keyspace to another keyspace.
Also can run below command to move all the keys from keyspace 1 to keyspace 0 :
Verify import status, using below commands : (inside redis-cluster-util container)
redis-cli -h 10.0.1.90-p 6379 info keyspace
It will run Redis info command on node 10.0.1.90:6379 and fetch keyspace information, like below:
# Keyspace db0:keys=33283,expires=0,avg_ttl=0
Now reshard all the slots to all instances evenly
The reshard command will again list the existing nodes, their IDs and the assigned slots.
redis-trib.rb reshard 10.0.1.90:6379
Parameters:
host:port of any node from the cluster.
Output:
It will prompt for number of slots to move – here (16384 /3 Masters = 5461)
Receiving node id – here id of master node 2
Source node IDs – id of first instance which has currently all the slots. (master 1)
And prompt to proceed – press ‘yes’
Repeat above step and for receiving node id, give id of master node 3.
After the above step, all 3 masters will have equal slots and imported keys will be distributed among the master nodes.
Put keys to cluster for verification
redis-cli -h 10.0.1.90-p 6379 set foo barOKredis-cli -h 10.0.1.90-p 6379 set foo bar(error) MOVED481310.0.9.203:6379
Above error shows that server saved this key to instance 10.0.9.203:6379, so client redirected it. To follow redirection, use flag “-c” which says it is a cluster mode, like:
redis-cli -h 10.0.1.90-p 6379-c set foo barOK
Redis Entrypoint
Application entrypoint for Redis cluster is mostly depends how your Redis client handles cluster support. Generally connecting to one of master nodes should do the work.
Use below host:port in your applications :
redis.marathon.l4lb.thisdcos.directory:6379
Automation of Redis Cluster Creation
We have automation script in place to deploy 6 node Redis cluster and form a cluster between them.
It deploys 6 marathon apps for 6 Redis nodes. All nodes are deployed on different nodes with CLUSTER_NAME as prefix to volume name.
Once all nodes are up and running, it deploys redis-cluster-util app which will be used to form Redis cluster.
Then it will print the Redis nodes and their IP addresses and prompt the user to proceed cluster creation.
If user selects to proceed, it will run redis-cluster-util app and create the cluster using IP addresses collected. Util container will prompt for some input that the user has to select.
Conclusion
We learned about Redis cluster deployment on DCOS with Persistent storage using RexRay. We also learned how rexray automatically manages volumes over aws ebs and how to integrate them in DCOS apps/services. We saw how to use redis-cluster-util container to manage Redis cluster for different purposes, like forming cluster, resharding, importing existing dump.rdb data etc. Finally, we looked at the automation part of whole cluster setup using dcos cli and bash.