Imagine handling an e-commerce site that generates data inflow at a very high rate. It is challenging to manage data traffic because the volumes of data are produced in varied formats or structures. DynamoDB is a NoSQL database that helps you scale your data effortlessly according to your workloads. It enables you to track your resource usage and automatically adjust it according to the traffic inflow or outflow.

For effective resource management, you also need to understand your data’s uniqueness. Databricks help you understand the data schemas using generative AI and natural language processing. Integrating data from DynamoDB to Databricks allows you to analyze data efficiently and discover strategies to enhance operational efficiency.

In this article, you will explore how to move data from DynamoDB to Databricks using different methods.

Why Integrate Data from DynamoDB to Databricks? 

There are several benefits of connecting DynamoDB to Dataricks. Here are a few of them: 

  • Enhance Data Synchronization: Databricks combines the strengths of a data warehouse and a data lake. You can build a data lakehouse to improve data synchronization. It provides timely access to your enterprise data and reduces the complexities of handling different data systems. 
  • Build Large Language Models: You can load DynamoDB file to Databricks and build customized large language models (LLM) for the data stored in the file. Databricks supports open-source tools such as DeepSpeed that help you make an LLM foundation. This allows you to train your data to generate a more accurate result related to your query.
  • Streaming Analytics: Databricks’s Apache Spark structured streaming enables you to work with streaming data. It also allows you to integrate with Delta Lake, which helps you with incremental data loading.

Overview of Amazon DynamoDB 

Amazon DynamoDB is a NoSQL database service that builds modern, serverless applications. Unlike most relational databases, DynamoDB stores data in key-value pairs, where the key represents the unique identifier, enabling faster queries. It offers automated horizontal scaling, which allows you to manage your resources according to workload or data traffic. 

For applications that require complex logic, DynamoDB provides server-side support. It simplifies the transactions and coordinates data across multiple sources within and across the tables. The platform also offers point-in-time recovery to protect your business data from accidental deletes or write operations. Amazon DynamoDB provides a fully managed experience, helping you run high-performance applications with limitless scalability.

Overview of Databricks

Databricks is a well-known data intelligence platform that combines the strengths of AI and Lakehouse to help you maintain and analyze your enterprise-grade data. It is built on the Apache Spark framework, incorporating the functionalities of a data lakehouse and connecting different sources to a single platform. 

Databricks’ collaborative workspace allows you to share data between teams in near real-time and develop and implement models using business intelligence (BI) and generative AI tools. With Databricks, you get quality, control, and tools that help you achieve breakthrough results by analyzing your business data.

Methods to Connect DynamoDB to Databricks

You can use the below-mentioned methods to learn how to load data from DynamoDB to Databricks.

Method 1: Integrating Data from DynamoDB to Databricks Using S3 Bucket

In this method, you will learn how to insert DynamoDB data into Databricks table using an S3 Bucket.

Step 1: Export Data from DynamoDB into an S3 Bucket

DynamoDB supports two types of exports. The first is full export, where you can take snapshots of your data at any point in time. The second is incremental export, where you can extract data from the DynamoDB table that was changed, deleted, or updated during a specific period. 


DynamoDB allows you to export data to an S3 Bucket using AWS Management Console and AWS CLI.

Follow the steps to export data from the DynamoDB table to an S3 Bucket using Amazon CLI:

  • Enable point-in-time-recovery for the table you want to export using the following command:
aws dynamodb update-continuous-backups \
    --table-name tablename \
    --point-in-time-recovery-specification PointInTimeRecoveryEnabled=True
  • You can implement export using either of the following ways:
    • To implement full export, execute the following command:
aws dynamodb export-table-to-point-in-time \
  --table-arn arn:aws:dynamodb:us-west-2:123456789012:table/MusicCollection \
  --s3-bucket ddb-export-tablename-9012345678 \
  --s3-prefix addprefix \
  --export-format DYNAMODB_JSON \
  --export-time 1604632434 \
  --s3-bucket-owner 9012345678 \
  --s3-sse-algorithm AES256
  • To implement the incremental export, execute the following command:
aws dynamodb export-table-to-point-in-time \
  --table-arn arn:aws:dynamodb:REGION:ACCOUNT:table/TABLENAME \
  --s3-bucket BUCKET --s3-prefix PREFIX \
  --incremental-export-specification ExportFromTime=1693569600,ExportToTime=1693656000,ExportViewType=NEW_AND_OLD_IMAGES \
  --export-type INCREMENTAL_EXPORT

Step 2: Import Data into Databricks from S3 Bucket

You can use the COPY INTO command which loads the data from an AWS S3 Bucket into your Databricks table. 


  • Configure your Databricks SQL warehouse to run commands for loading data from the S3 Bucket. 
  • Configure the Unity Catalog or external location using the storage credentials. 
  • You need to confirm that access to a cloud storage system can read the data in source files. 
  • You need to create a table where you will store the extracted data. 

Follow the steps to load data to your Databricks table: 

  • Go to the side menu and click on Create>Query. 
  • In the SQL editor menu bar, select running SQL warehouse. 
  • Run the following code in the SQL editor: 
COPY INTO <catalog-name>.<schema-name>.<table-name>
FROM 's3://<s3-bucket>/<folder>/'
  'header' = 'true',
  'inferSchema' = 'true'
   'mergeSchema' = 'true'
SELECT * FROM <catalog_name>.<schema_name>.<table_name>;

Note: Replace <s3-bucket> with your S3 Bucket name and <folder> with the name of the folder in which your S3 Bucket is present. 

Limitations for Integrating Data from DynamoDB to Databricks Using S3 Bucket

  • The S3 Bucket offers eventual consistency to overwrite PUTS and DELETEs, which can lead to a delay in reflecting changes in the data after it gets updated. 
  • Using the COPY INTO command simultaneously to write data to a table can cause confusion when handling large-scale data processing.

Method 2: Integrating Data from DynamoDB to Databricks Using Hevo

Hevo is a real-time ELT platform that lets you easily load DynamoDB file to Databricks. You can integrate data between the two platforms using Hevo’s no-code, flexible, and automated data pipeline. It connects your source to your destination and prepares your data for analysis. Hevo also offers 150+ data sources, including databases, data warehouses, cloud platforms, and more, from which you can transfer data to your desired destination.

Benefits of Using Hevo:

  • Automated Schema Mapping: This feature of Hevo automatically reads the source data’s schema and replicates it to your destination. It saves time and increases data efficiency. 
  • Data Transformation: To avoid inconsistencies and errors in your data, you need to clean it before performing data ingestion. You can use Hevo’s Python-based or drag-drop transformation to prepare your data for analysis. 
  • Incremental Data Loading: Hevo lets you load only modified or updated source data to your destination. This helps you save time and avoid duplicating or overwriting the same data again. 

Here’s how Hevo helps you connect DynamoDB to Databricks in two simple steps.

Step 1: Configure Amazon DynamoDB as Your Source

To support Change Data Capture (CDC), Hevo uses DynamoDB data streams, which are time-order sequences of item-level changes in the DynamoDB table. 

To keep track of the data read from the data stream, Hevo uses two ways of replicating data to manage data ingestion:

  • Amazon Kinesis Data Streams 
  • Amazon DynamoDB Streams


Follow the steps to configure your source settings: 

  • Go to PIPELINES in the Navigation Bar. 
  • Click on +CREATE in the Pipeline View List. 
  • Search for and select DynamoDB as your source on the Select Source Type page.
  • On the Configure your DynamoDB source page, specify the mandatory details to set the connection.
  • Once you have filled in all the details, click TEST & CONTINUE.
DynamoDB to Databricks: Configure Source Settings
Configure Source Settings

Refer to Hevo documentation for more information on configuring DynamoDB as a source.

Step 2: Configure Databricks as Your Destination 

Hevo supports Databricks warehouses hosted on AWS, Azure, or GCP platforms. You can use the Databricks Partner Connect Method to configure Databricks as your destination.


Follow the steps to configure your destination settings: 

  • Click on DESTINATIONS in the Navigation Bar. 
  • Click on +CREATE in the Destination List View.
  • Search for and select Datarbricks on the Add Destination page. 
  • On the Configure your Databricks Destination page, specify the mandatory details to set up the connection. Once you have filled in all the details, click Test Connection and SAVE & CONTINUE.
DynamoDB to Databricks: Configure Destination Settings
Configure Destination Settings

Refer to Hevo documentation for more information on configuring Databricks as a destination.

Get started for Free with Hevo!

Use Cases of DynamoDB to Databricks Integration

  • Identifying Potential Cyberthreats: Databricks helps you identify potential cyber threats using real-time data analysis. It allows you to monitor your website’s network traffic and identify suspicious activities by enabling a proactive threat detection model that protects sensitive data.
  • Preventing churn: When you ingest your customer data from DynamoDB to Databricks, you can use its machine learning capabilities to build predictive models. These models help you identify your customers’ pain points and prevent churns. 
  • Knowledge Sharing: Using Databricks’ collaborative workspace, you can share your data, combine codes, and query it interactively. This allows you to enhance your data visualization and increase operational efficiency. 
Are you looking for an easy way to migrate data from DynamoDBto Databricks? Solve your data replication problems with Hevo Data’s reliable, no-code, automated pipelines with 150+ connectors.
Get your free trial right away!


Integrating data from DynamoDB to Databricks enables you to process and analyze collaboratively at a large scale. You can share insights, build machine learning models, and extract insights by using those models to make more informed decisions about your business activities. There are two ways you can integrate data between DyanamoDB and Databricks. You can use S3 Bucket for data integration, which can be a lengthy approach. You can also use Hevo, which automates your data indigestion process with its flexible data pipeline. 

FAQs (Frequently Asked Questions)

Q. How can you capture change data in DynamoDb and write it in the Delta table in Databricks?

Follow the steps to write a change data capture from Dynamodb to Delta table in Databricks: 

  • Enable the datastreams in DynamoDB using AWS CLI or AWS SDK.
  • Create a Lambda function to send updates from DynamoDB to external locations. 
  • Export the required libraries to your Databricks workspace and define a schema to view data from DynamoDB.
  • Create a Delta table to store all of the data changes.
Customer Experience Engineer, Hevo Data

Skand, with 2.5 years of experience, specializes in MySQL, Postgres, and REST APIs at Hevo. He efficiently troubleshoots customer issues, contributes to knowledge base and SOPs, and assists customers in achieving their use cases through Hevo's platform.

All your customer data in one place.