Skip to content

Failover with Jedis

Jedis supports failover for your Redis deployments. This is useful when: 1. You have more than one Redis deployment. This might include two independent Redis servers or two or more Redis databases replicated across multiple active-active Redis Enterprise clusters. 2. You want your application to connect to and use one deployment at a time. 3. You want your application to fail over to the next available deployment if the current deployment becomes unavailable.

Jedis will fail over to a subsequent Redis deployment after reaching a configurable failure threshold. This failure threshold is implemented using a circuit breaker pattern.

You can also configure Jedis to retry failed calls to Redis. Once a maximum number of retries have been exhausted, the circuit breaker will record a failure. When the circuit breaker reaches its failure threshold, a failover will be triggered on the subsequent operation.

The remainder of this guide describes:

  • A basic failover configuration
  • Supported retry and circuit breaker settings
  • Failback and the cluster selection API

We recommend that you read this guide carefully and understand the configuration settings before enabling Jedis failover in production.

Basic usage

To configure Jedis for failover, you specify an ordered list of Redis databases. By default, Jedis will connect to the first Redis database in the list. If the first database becomes unavailable, Jedis will attempt to connect to the next database in the list, and so on.

Suppose you run two Redis deployments. We'll call them redis-east and redis-west. You want your application to first connect to redis-east. If redis-east becomes unavailable, you want your application to connect to redis-west.

Let's look at one way of configuring Jedis for this scenario.

First, create an array of ClusterConfig objects, one for each Redis database.

JedisClientConfig config = DefaultJedisClientConfig.builder().user("cache").password("secret").build();

ClusterConfig[] clientConfigs = new ClusterConfig[2];
clientConfigs[0] = new ClusterConfig(new HostAndPort("redis-east.example.com", 14000), config);
clientConfigs[1] = new ClusterConfig(new HostAndPort("redis-west.example.com", 14000), config);

The configuration above represents your two Redis deployments: redis-east and redis-west. You'll use this array of configuration objects to create a connection provider that supports failover.

Use the MultiClusterClientConfig builder to set your preferred retry and failover configuration, passing in the client configs you just created. Then build a MultiClusterPooledConnectionProvider.

MultiClusterClientConfig.Builder builder = new MultiClusterClientConfig.Builder(clientConfigs);
builder.circuitBreakerSlidingWindowSize(10);
builder.circuitBreakerSlidingWindowMinCalls(1);
builder.circuitBreakerFailureRateThreshold(50.0f);

MultiClusterPooledConnectionProvider provider = new MultiClusterPooledConnectionProvider(builder.build());

Internally, the connection provider uses a highly configurable circuit breaker and retry implementation to determine when to fail over. In the configuration here, we've set a sliding window size of 10 and a failure rate threshold of 50%. This means that a failover will be triggered if 5 out of any 10 calls to Redis fail.

Once you've configured and created a MultiClusterPooledConnectionProvider, instantiate a UnifiedJedis instance for your application, passing in the provider you just created:

UnifiedJedis jedis = new UnifiedJedis(provider);

You can now use this UnifiedJedis instance, and the connection management and failover will be handled transparently.

Configuration options

Under the hood, Jedis' failover support relies on resilience4j, a fault-tolerance library that implements retry and circuit breakers.

Once you configure Jedis for failover using the MultiClusterPooledConnectionProvider, each call to Redis is decorated with a resilience4j retry and circuit breaker.

By default, any call that throws a JedisConnectionException will be retried up to 3 times. If the call continues to fail after the maximum number of retry attempts, then the circuit breaker will record a failure.

The circuit breaker maintains a record of failures in a sliding window data structure. If the failure rate reaches a configured threshold (e.g., when 50% of the last 10 calls have failed), then the circuit breaker's state transitions from CLOSED to OPEN. When this occurs, Jedis will attempt to connect to the next Redis database in its client configuration list.

The supported retry and circuit breaker settings, and their default values, are described below. You can configure any of these settings using the MultiClusterClientConfig.Builder builder. Refer the basic usage above for an example of this.

Retry configuration

Jedis uses the following retry settings:

Setting Default value Description
Max retry attempts 3 Maximum number of retry attempts (including the initial call)
Retry wait duration 500 ms Number of milliseconds to wait between retry attempts
Wait duration backoff multiplier 2 Exponential backoff factor multiplied against wait duration between retries. For example, with a wait duration of 1 second and a multiplier of 2, the retries would occur after 1s, 2s, 4s, 8s, 16s, and so on.
Retry included exception list [JedisConnectionException] A list of Throwable classes that count as failures and should be retried.
Retry ignored exception list null A list of Throwable classes to explicitly ignore for the purposes of retry.

To disable retry, set maxRetryAttempts to 1.

Circuit breaker configuration

Jedis uses the following circuit breaker settings:

Setting Default value Description
Sliding window type COUNT_BASED The type of sliding window used to record the outcome of calls. Options are COUNT_BASED and TIME_BASED.
Sliding window size 100 The size of the sliding window. Units depend on sliding window type. When COUNT_BASED, the size represents number of calls. When TIME_BASED, the size represents seconds.
Sliding window min calls 100 Minimum number of calls required (per sliding window period) before the CircuitBreaker will start calculating the error rate or slow call rate.
Failure rate threshold 50.0f Percentage of calls within the sliding window that must fail before the circuit breaker transitions to the OPEN state.
Slow call duration threshold 60000 ms Duration threshold above which calls are classified as slow and added to the sliding window.
Slow call rate threshold 100.0f Percentage of calls within the sliding window that exceed the slow call duration threshold before circuit breaker transitions to the OPEN state.
Circuit breaker included exception list [JedisConnectionException] A list of Throwable classes that count as failures and add to the failure rate.
Circuit breaker ignored exception list null A list of Throwable classes to explicitly ignore for failure rate calculations.

Fallback configuration

Jedis uses the following fallback settings:

Setting Default value Description
Fallback exception list [CallNotPermittedException, JedisConnectionException] A list of Throwable classes that trigger fallback.

Failover callbacks

In the event that Jedis fails over, you may wish to take some action. This might include logging a warning, recording a metric, or externally persisting the cluster connection state, to name just a few examples. For this reason, MultiPooledConnectionProvider lets you register a custom callback that will be called whenever Jedis fails over to a new cluster.

To use this feature, you'll need to design a class that implements java.util.function.Consumer. This class must implement the accept method, as you can see below.

import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.function.Consumer;

public class FailoverReporter implements Consumer<String> {

    @Override
    public void accept(String clusterName) {
        Logger logger = LoggerFactory.getLogger(FailoverReporter.class);
        logger.warn("Jedis failover to cluster: " + clusterName);
    }
}

You can then pass an instance of this class to your MultiPooledConnectionProvider.

FailoverReporter reporter = new FailoverReporter();
provider.setClusterFailoverPostProcessor(reporter);

The provider will call your accept whenever a faoliver occurs.

Failing back

We believe that failback should not be automatic. If Jedis fails over to a new cluster, Jedis will not automatically fail back to the cluster that it was previously connected to. This design prevents a scenario in which Jedis fails back to a cluster that may not be entirely healthy yet.

That said, we do provide an API that you can use to implement automated failback when this is appropriate for your application.

Failback scenario

When a failover is triggered, Jedis will attempt to connect to the next Redis server in the list of server configurations you provide at setup.

For example, recall the redis-east and redis-west deployments from the basic usage example above. Jedis will attempt to connect to redis-east first. If redis-east becomes unavailable (and the circuit breaker transitions), then Jedis will attempt to use redis-west.

Now suppose that redis-east eventually comes back online. You will likely want to fail your application back to redis-east. However, Jedis will not fail back to redis-east automatically.

In this case, we recommend that you first ensure that your redis-east deployment is healthy before you fail back your application.

Failback behavior and cluster selection API

Once you've determined that it's safe to fail back to a previously-unavailable cluster, you need to decide how to trigger the failback. There are two ways to accomplish this:

  1. Use the cluster selection API
  2. Restart your application

Fail back using the cluster selection API

MultiClusterPooledConnectionProvider exposes a method that you can use to manually select which cluster Jedis should use. To select a different cluster to use, pass the cluster's numeric index to setActiveMultiClusterIndex().

The cluster's index is a 1-based index derived from its position in the client configuration. For example, suppose you configure Jedis with the following client configs:

ClusterConfig[] clientConfigs = new ClusterConfig[2];
clientConfigs[0] = new ClusterConfig(new HostAndPort("redis-east.example.com", 14000), config);
clientConfigs[1] = new ClusterConfig(new HostAndPort("redis-west.example.com", 14000), config);

In this case, redis-east will have an index of 1, and redis-west will have an index of 2. To select and fail back to redis-east, you would call the function like so:

provider.setActiveMultiClusterIndex(1);

This method is thread-safe.

If you decide to implement manual failback, you will need a way for external systems to trigger this method in your application. For example, if your application exposes a REST API, you might consider creating a REST endpoint to call setActiveMultiClusterIndex and fail back the application.

Fail back by restarting the application

When your application starts, Jedis will attempt to connect to each cluster in the order that the clusters appear in your client configuration. It's important to understand this, especially in the case where Jedis has failed over. If Jedis has failed over to a new cluster, then restarting the application may result in an inadvertent failback. This can happen only if a failed cluster comes back online and the application subsequently restarts.

If you need to avoid this scenario, consider using a failover callback, as described above, to externally record the name of the cluster that your application was most recently connected to. You can then check this state on startup to ensure that you application only connects to the most recently used cluster. For assistance with this technique, start a discussion.