DEV Community


New Features in Amazon DynamoDB - PartiQL, Export to S3, Integration with Kinesis Data Streams

Avid learner of technology solutions around databases, big-data, Machine Learning. 5x AWS Certified | 5x Oracle Certified. Connect on Twitter @anandp86
・12 min read

Every time with AWS re:Invent around, AWS releases many new features over a period of month. In this blog post I will touch on 3 new features which were introduced for Amazon DynamoDB. DynamoDB is a non-relational managed database with single digit millisecond performance at any scale.

New Features in Amazon DynamoDB -

  1. PartiQL - SQL-compatible query language for Amazon DynamoDB.
  2. Export to S3 - Export Amazon DynamoDB table to S3. In this blog I have added a use-case of deserializing the DynamoDB items, writing it to S3 and query using Athena.
  3. Direct integration of DynamoDB with Kinesis Streams - Stream item-level images of Amazon DynamoDB as a Kinesis Data Stream.

To start with, lets look at the new Amazon DynamoDB console. I have 2 DDB tables to play around with for this blog. The Books table is partitioned by Author and sorted by Title. The Movies table is partitioned by year and sorted by title.

Lets jump on to the features:

  1. PartiQL - You can use SQL to select, insert, update and delete items from Amazon DynamoDB. Currently you can use PartiQL for DynamoDB from the Amazon DynamoDB console, the AWS Command Line Interface (AWS CLI), and DynamoDB APIs. For this blog, I am using the AWS console.

DynamoDB > PartiQL editor


Simple select SQL

SELECT * FROM Books where Author='William Shakespeare'
Title Formats Author Category
Hamlet { "Hardcover" : { "S" : "GVJZQ7JK" }, "Paperback" : { "S" : "A4TFUR98" }, "Audiobook" : { "S" : "XWMGHW96" } } William Shakespeare Drama

The following SQL returns the title, hardcover and category using key path -

SELECT Title, Formats['Hardcover'], category FROM Books where Author='John Grisham'
Category Title Hardcover
Suspense The Firm Q7QWE3U2
Suspense The Rainmaker J4SUKVGU
Thriller The Reckoning null

The following SQL uses "contains" function which returns TRUE if attribute category has string 'Suspense' -

SELECT Title, Formats['Audiobook'], Category FROM Books where Author='John Grisham' and contains(Category, 'Suspense')
year title release_date rank
2011 Sherlock Holmes: A Game of Shadows 2011-12-10T00:00:00Z 570


Insert a single item -

INSERT INTO Books value {'Title' : 'A time to kill', 'Author' : 'John Grisham', 'Category' : 'Suspense' }

SELECT * FROM Books WHERE Title='A time to kill'
Author Title Category
John Grisham A time to kill Suspense

"INSERT INTO SELECT" SQL fails with ValidationException: Unsupported operation: Inserting multiple items in a single statement is not supported, use "INSERT INTO tableName VALUE item" instead


In the previous insert sql, Formats column was null. So lets update the Formats column for the book.

UPDATE Books SET Formats={'Hardcover':'J4SUKVGU' ,'Paperback': 'D7YF4FCX'} WHERE Author='John Grisham' and Title='A time to kill'
Title Formats Author Category
A time to kill {"Hardcover":{"S":"J4SUKVGU"},"Paperback":{"S":"D7YF4FCX"}} John Grisham Suspense

You can use update sql to remove key from map -

UPDATE Books REMOVE Formats.Paperback WHERE Author='John Grisham' and Title='A time to kill'
Title Formats Author Category
A time to kill {"Hardcover":{"S":"J4SUKVGU"}} John Grisham Suspense


DELETE FROM Books WHERE Author='John Grisham' and Title='A time to kill'

For more references -

2. EXPORT TO S3 - Export full Amazon DynamoDB table to Amazon S3 bucket. Using DynamoDB table export, you can export data from an Amazon DynamoDB table from any time within your point-in-time recovery window. To do so the table must have point in time recovery (PITR) enabled. If PITR is not enabled for the table, the Export to S3 will report error asking it to be enabled.

DynamoDB > Exports to S3

After the export is complete it generates a manifest-summary.json file summarizing the export details and a manifest-file.json containing the details of S3 file locations.

Further to extend on Export to S3 feature to perform analytics and complex queries on your data you can create a data pipeline to deserialize the DynamoDB JSON format and query data from AWS Athena.

The workflow contains the following steps:

  • Time based CloudWatch event is triggered.
  • This event triggers AWS Lambda function.
  • DynamoDB Export to S3 is initiated.
  • It writes the data in DynamoDB JSON format to S3 RAW bucket. The S3 objects are zipped json files.
  • For each .json.gz file in S3 RAW bucket an event notification is set to trigger AWS Lambda. The detail is show below in S3 Event to trigger AWS Lambda section.
  • AWS Lambda reads the S3 object and deserializes the DynamoDB JSON format data using DynamoDB TypeDeserializer class. This class deserializes DynamoDB types to Python types. The deserialized data is written to S3 Content bucket in Parquet format. Code is in Lambda function code section.
  • AWS Lambda function updates the table location in AWS Glue catalog.
  • Query the data using AWS Athena.

S3 Event to trigger AWS Lambda

Lambda function code

import io
import gzip
import json
import boto3
import uuid
import pandas as pd
import awswrangler as wr
from datetime import datetime
from urllib.parse import unquote_plus

def update_glue_table(*, database, table_name, new_location, region_name):
    """ Update AWS Glue non-partitioned table location

    glue = boto3.client("glue", region_name=region_name)

    response = glue.get_table(
        DatabaseName=database, Name=table_name)

    table_input = response["Table"]
    current_location = table_input["StorageDescriptor"]["Location"]

    table_input.pop("UpdateTime", None)
    table_input.pop("CreatedBy", None)
    table_input.pop("CreateTime", None)
    table_input.pop("DatabaseName", None)
    table_input.pop("IsRegisteredWithLakeFormation", None)
    table_input.pop("CatalogId", None)

    table_input["StorageDescriptor"]["Location"] = new_location

    response = glue.update_table(

    return response

def lambda_handler(event, context): 
    Uses class TypeDeserializer which deserializes DynamoDB types to Python types 
    Example - 

    raw data format :
        [{'ACTIVE': {'BOOL': True}, 'params': {'M': {'customer': {'S': 'TEST'}, 'index': {'N': '1'}}}}, ]
    deserialized data format:
        [{'ACTIVE': True, 'params': {'customer': 'TEST', 'index': Decimal('1')}}]

    s3client = boto3.client('s3')
    athena_db = "default"
    athena_table = "movies"
    for record in event['Records']:
        bucket = record['s3']['bucket']['name']
        key = unquote_plus(record['s3']['object']['key'])
        response = s3client.get_object(Bucket=bucket, Key=key)
        content = response['Body'].read()
        with gzip.GzipFile(fileobj=io.BytesIO(content), mode='rb') as fh:
            data = [json.loads(line) for line in fh]
    all_data = []
    deserializer = boto3.dynamodb.types.TypeDeserializer()
    for row in data:
        all_data.append({k: deserializer.deserialize(v) for k,v in row['Item'].items()})

    data_df = pd.DataFrame(all_data)
    dt = datetime.utcnow().strftime("%Y-%m-%d-%H-%M")
    s3_path="s3://%s/dynamodb/%s/content/dt=%s/" % (bucket, athena_table, dt)
        dataset = True,

    update_response = update_glue_table(

    if update_response["ResponseMetadata"]["HTTPStatusCode"] == 200:
        return (f"Successfully updated glue table location - {athena_db}.{athena_table}")
        return (f"Failed updating glue table location - {athena_db}.{athena_table}")

Query data from Athena

SELECT title, info.actors, info.rating, info.release_date, year FROM movies where title='Christmas Vacation'
title actors rating release_date year
1 Christmas Vacation [Chevy Chase, Beverly D'Angelo, Juliette Lewis] 0.73 1989-11-30T00:00:00Z 1989

For more references -

3. DynamoDB integration with Kinesis Stream.

DynamoDB Streams captures a time-ordered sequence of item-level modifications in a DynamoDB table. Earlier to publish DynamoDB data to S3 in near real time, one of the ways was to enable DynamoDB streams and use AWS Lambda function to forward data to Kinesis Firehose which published the data to S3. To do so you can use a handy package provided by aws labs - .

With several such use cases, AWS has now integrated AWS DynamoDB directly with Amazon Kinesis Stream. Now you can capture item-level changes in your DynamoDB tables as a Kinesis data stream. This can enable you to publish the data to S3 as shown in the below pipeline.

To setup the pipeline, create a Amazon Kinesis Data Stream.

After setting up the Amazon Kinesis Data Stream, create Amazon Kinesis Firehose. The source to Kinesis Firehose will be Amazon Kinesis Data Stream and the destination will be S3.

To proceed move on to enable streaming to Kinesis.

DynamoDB > Table > Kinesis data stream details > Manage streaming to Kinesis

Once the stream is enabled any item-level change in the table will be captured and written to Amazon S3 bucket. Below is an example of the record which was updated in DynamoDB using PartiQL. The record contains approximate creation date time of the record in DynamoDB streams, along with New and Old image of the record. These records can be parsed using AWS Lambda or AWS Glue and stored in Data Lake for analytical use-cases.

   "awsRegion": "us-west-2",
   "dynamodb": {
     "ApproximateCreationDateTime": 1606714671542,
     "Keys": {
       "Author": {
         "S": "James Patterson"
       "Title": {
         "S": "The President Is Missing"
     "NewImage": {
       "Title": {
         "S": "The President Is Missing"
       "Formats": {
         "M": {
           "Hardcover": {
             "S": "JSU4KGVU"
       "Author": {
         "S": "James Patterson"
       "Category": {
         "S": "Mystery"
     "OldImage": {
       "Title": {
         "S": "The President Is Missing"
       "Formats": {
         "M": {
           "Hardcover": {
             "S": "JSU4KGVU"
           "Paperback": {
             "S": "DY7F4CFX"
       "Author": {
         "S": "James Patterson"
       "Category": {
         "S": "Mystery"
     "SizeBytes": 254
   "eventID": "bcaaf073-7e0d-49c2-818e-fe3cf7e5f18a",
   "eventName": "MODIFY",
   "userIdentity": null,
   "recordFormat": "application/json",
   "tableName": "Books",
   "eventSource": "aws:dynamodb"

To conclude, in this post I have introduced you to PartiQL which provides SQL-compatible query for Amazon DynamoDB. We also looked at the Export to S3 feature and how we can create an end to end pipeline to query the Amazon DynamoDB data in Amazon S3 bucket using AWS Athena. And finally we looked at real time analytics use-case where you can enable streams to capture item-level changes in Amazon DynamoDB table as Kinesis data streams.

Discussion (0)