DEV Community

Abhishek Gupta for Microsoft Azure

Posted on • Originally published at devblogs.microsoft.com

Build fault tolerant applications with Cassandra API for Azure Cosmos DB

Azure Cosmos DB is a resource governed system that allows you to execute a certain number of operations per second based on the provisioned throughput you have configured. If clients exceed that limit and consume more request units than what was provisioned, it leads to rate limiting of subsequent requests and exceptions being thrown – they are also referred to as 429 errors.

With the help of a practical example, I'll demonstrate how to incorporate fault-tolerance in your Go applications by handling and retrying operations affected by these rate limiting errors. To help you follow along, the sample application code for this blog is available on GitHub - it uses the gocql driver for Apache Cassandra.

In this post, we'll go through:

  • Initial setup and configuration before running the sample application
  • Execution of various load test scenarios and analyze the results
  • A quick overview of the Retry Policy implementation.

One way of tackling rate limiting is by adjusting provisioned throughput to meet your application requirements. There are multiple ways to do this, including using Azure portal, Azure CLI, and CQL (Cassandra Query Language) commands.

But, what if you wanted to handle these errors in the application itself?

The good thing is that the Cassandra API for Azure Cosmos DB translates the rate limiting exceptions into overloaded errors on the Cassandra native protocol. Since the gocql driver allows you to plugin your own RetryPolicy, you can write a custom implementation to intercept these errors and retry them after a certain (cool down) time period. This policy can then be applied to each Query or at a global level using a ClusterConfig.

The Azure Cosmos DB extension library makes it quite easy to use Retry Policies in your Java applications. An equivalent Go version is available on GitHub and has been used in the sample application for this blog post.

Retry Policy in action

As promised, you will walk through the entire process using a simple yet practical example. The sample application used to demonstrate the concepts is a service that exposes a REST endpoint to POST orders data which is persisted to a Cassandra table in Azure Cosmos DB.

You will run a few load tests on this API service to see how rate limiting manifests itself and how it's handled.

Pre-requisites

Start by installing hey, a load testing program. You can download OS specific binaries (64-bit) for Linux, Mac and Windows (please refer to the GitHub repo for latest information in case you face issues downloading the utility)

You can use any other tool that allows you to generate load on an HTTP endpoint

Clone this GitHub repo and change into the right directory:

git clone github.com/abhirockzz/cosmos-go-rate-limiting 
cd cosmos-go-rate-limiting
Enter fullscreen mode Exit fullscreen mode

Setup Azure Cosmos DB

Create an Azure Cosmos DB account with the Cassandra API option selected

To create a Keyspace and Table, use the following CQL:

    CREATE KEYSPACE ordersapp WITH REPLICATION = {'class' : 'SimpleStrategy'};

    CREATE TABLE ordersapp.orders (
        id uuid PRIMARY KEY,
        amount int,
        state text,
        time timestamp
    );
Enter fullscreen mode Exit fullscreen mode

Start the application

Open a terminal and set the environment variables for the application:

export COSMOSDB_CASSANDRA_CONTACT_POINT=.cassandra.cosmos.azure.com 
export COSMOSDB_CASSANDRA_PORT=10350 
export COSMOSDB_CASSANDRA_USER= 
export COSMOSDB_CASSANDRA_PASSWORD= 
#optional (default: 5) 
#export MAX_RETRIES=
Enter fullscreen mode Exit fullscreen mode

To start the application:

go run main.go 

//wait for this output 
Connected to Azure Cosmos DB
Enter fullscreen mode Exit fullscreen mode

To test whether the application is working as expected, insert a few orders by invoking the REST endpoint (once for each order) from a different terminal:

curl http://localhost:8080/orders
Enter fullscreen mode Exit fullscreen mode

The application generates random data so you don't have to enter it while invoking the endpoint

Confirm that the order was successfully stored. You can use the hosted CQL shell in the Azure portal and execute the below query:

    select count(*) from ordersapp.orders;

    // you should see this output
    system.count(*) 
    ----------------- 
        1 
    (1 rows)
Enter fullscreen mode Exit fullscreen mode

You're all set.

Let the load tests begin!

Invoke the REST endpoint with 300 requests. This is enough to overload the system since you only have 400 RU/s allocated by default.

To start the load test:

hey -t 0 -n 300 http://localhost:8080/orders
Enter fullscreen mode Exit fullscreen mode

Notice the logs in the application terminal. In the beginning, you will see that the orders are being successfully created. For example:

Added order ID 25a8cec1-e67a-11ea-9c17-7f242c2eeac0
Added order ID 25a8f5ef-e67a-11ea-9c17-7f242c2eeac0
Added order ID 25a8f5ea-e67a-11ea-9c17-7f242c2eeac0
...
Enter fullscreen mode Exit fullscreen mode

After a while, as the throughput degrades and eventually exceeds the provisioned limit, Azure Cosmos DB will rate limit the application requests. This will manifest itself in the form of an error which looks similar to this:

    Request rate is large: ActivityID=ac78fac3-5c36-4a20-8ad7-4b2d0768ffe4, RetryAfterMs=112, Additional details='Response status code does not indicate success: TooManyRequests (429); Substatus: 3200; ActivityId: ac78fac3-5c36-4a20-8ad7-4b2d0768ffe4; Reason: ({
      "Errors": [
        "Request rate is large. More Request Units may be needed, so no changes were made. Please retry this request later. Learn more: http://aka.ms/cosmosdb-error-429"
      ]
    });
Enter fullscreen mode Exit fullscreen mode

In the error message above, notice the following: TooManyRequests (429) and RetryAfterMs=112

Observing Query errors

To keep things simple, we will use the log output for testing/diagnostic purposes. Any error (related to rate–limiting in this case) encountered during query execution is intercepted by a gocql.QueryObserver. The randomly generated order ID is also logged with each error message so that you can check the logs to confirm if the failed order has been re–tried and (eventually) stored in Azure Cosmos DB.

Here is the code snippet:

    ....
    type OrderInsertErrorLogger struct {
       orderID string
    }

    // implements gocql.QueryObserver
    func (l OrderInsertErrorLogger) ObserveQuery(ctx context.Context, oq gocql.ObservedQuery) {
      err := oq.Err
      if err != nil {
         log.Printf("Query error for order ID %sn%v", l.orderID, err)
      }
    }

    ....

    // the Observer is associated with each query
    rid, _ := uuid.GenerateUUID()
    err := cs.Query(insertQuery).Bind(rid, rand.Intn(200)+50, fixedLocation, time.Now()).Observer(OrderInsertErrorLogger{orderID: rid}).Exec()
    ....
Enter fullscreen mode Exit fullscreen mode

How many orders made it through?

Switch back to the load testing terminal and check some of the statistics (output has been redacted for brevity)

    Summary: 

      Total:        2.8507 secs 
      Slowest:      1.3437 secs 
      Fastest:      0.2428 secs 
      Average:      0.5389 secs 
      Requests/sec: 70.1592 
    .... 

    Status code distribution: 
      [200] 300 responses
Enter fullscreen mode Exit fullscreen mode

The numbers will differ in your specific case depending on multiple factors.

This is not a raw benchmarking test and neither do we have a production grade application, so you can ignore the  Requests/sec etc. But draw our attention to the Status code distribution  attribute which shows that our application responded with a HTTP 200  for all the requests.

Let's confirm the final numbers. Open the Cassandra Shell in the Azure Cosmos DB portal and execute the same query:

    select count(*) from ordersapp.orders;

    //output

    system.count(*)
    -----------------
        301
Enter fullscreen mode Exit fullscreen mode

You should see 300 additional rows (orders) have been inserted. The key takeaway is that all the orders were successfully stored in Azure Cosmos DB de–spite the rate limiting errors because our application code transparently retried them based on the Retry Policy that we configured (with a single line of code!)

clusterConfig.RetryPolicy = retry.NewCosmosRetryPolicy(numRetries)
Enter fullscreen mode Exit fullscreen mode

A note on dynamic throughput management

If your application spends most of its time operating at about 60-70% of it's throughput, using Autoscale provisioned throughput can help optimize your RU/s and cost usage by scaling down when not in use – you only pay for the resources that your workloads need on a per-hour basis.

So, what happens without the Retry Policy?

Deactivate the policy to see the difference

Stop the application (press control+c in the terminal), set an environment variable and re-start the application:

    export USE_RETRY_POLICY=false 
    go run main.go
Enter fullscreen mode Exit fullscreen mode

Before running the load test again, make a note of the number of rows in the orders table using  select count(*) from ordersapp.orders;

hey -t 0 -n 300 http://localhost:8080/orders
Enter fullscreen mode Exit fullscreen mode

In the application logs, you will notice the same rate limiting errors. In the terminal where you ran the load test, at the end of the output summary, you will see that some the requests failed to complete successfully i.e. they returned a response other than HTTP 200

    ...
    Status code distribution: 
      [200] 240 responses 
      [429] 60 responses
Enter fullscreen mode Exit fullscreen mode

Because the Retry Policy was not enforced, the application no longer re–tried the requests that failed due to rate-limiting.

Increase provisioned throughput

You can increase the Request Units using the Azure Portal (for example, double it to 800  RU/s) and run the same load test


hey -t 0 -n 300 http://localhost:8080/orders
Enter fullscreen mode Exit fullscreen mode

You will not see the rate limiting (HTTP429) errors now and relatively low numbers for latency, requests per second etc.

Try increasing the number of requests (use the -n flag) to see when the throughput threshold is breached for the application to get rate limited. As expected, all the orders will be persisted successfully (without any errors or retries)

The next section briefly covers how the custom Retry Policy works.

This is an experimental implementation, and you should write custom policies to suit fault-tolerance and performance requirements of your applications.

Behind the scenes

CosmosRetryPolicy adheres to the gocql.RetryPolicy interface by implementing the Attempt and GetRetry functions.

    type CosmosRetryPolicy struct {
        MaxRetryCount         int
        FixedBackOffTimeMs    int
        GrowingBackOffTimeMs  int
        numAttempts           int
    }
Enter fullscreen mode Exit fullscreen mode

Retry is initiated only if the number of retry attempts for that query are less than or equal to max retry config or max retry config is set to -1 (infinite retries)

    func (crp *CosmosRetryPolicy) Attempt(rq gocql.RetryableQuery) bool { 
        crp.numAttempts = rq.Attempts() 
        return rq.Attempts() <= crp.MaxRetryCount || crp.MaxRetryCount == -1
    }
Enter fullscreen mode Exit fullscreen mode

GetRetryType function detects the type of error and in the case or a rate-limited error (HTTP 429), it tries to extract the value for RetryAfterMs field (from the error message) and uses that to sleep before retrying the query.

    func (crp *CosmosRetryPolicy) GetRetryType(err error) gocql.RetryType {

       switch err.(type) {
       default:
             retryAfterMs := crp.getRetryAfterMs(err.Error())
             if retryAfterMs == -1 {
                 return gocql.Rethrow
             }
            time.Sleep(retryAfterMs)
            return gocql.Retry

    //other case statements have been omitted for brevity
    }
Enter fullscreen mode Exit fullscreen mode

Azure Cosmos DB provides you the flexibility to not only configure and adjust your throughput requirements using a variety of ways but also provides the basic primitive that allows applications to handle rate limiting errors, thereby making them robust and fault-tolerant. This blog post demonstrated how you can do this for Go applications, but the concepts are applicable to any language and its respective CQL compatible driver that you choose for working with the Cassandra API for Azure Cosmos DB.

To learn more:

Check out some of these resources from the official documentation:

Top comments (0)