Contact Us

May 16, 2025

May 13, 2026 3:46 pm

Optimizing Large Data Extension Imports in Salesforce Marketing Cloud

Share with

Understanding the Challenge

The original issue was that the team was trying to load 100 million records into a single data extension every day, but was experiencing timeout issues and slow processing times. They had split the data into four files, each with 25 million records, and were using an import activity to load the data into the data extension. However, this process was taking around 3.5 hours to complete, and they wanted to reduce the processing time to 1.5 hours.

Another approach they tried was to create a parallel automation to load the data into different data extensions and then merge them into a single data extension using a SQL activity. However, this approach was also resulting in timeout errors.

The root cause of the issue is that loading 100 million records into a single data extension is considered an extreme use case, and can result in performance issues. The data extension is not designed to handle such large volumes of data, and the import activity is not optimized for this type of workload.

Optimizing the Import Process

To optimize the import process, it is recommended to split the data into smaller, normalized tables, each with a subset of the attributes. This approach allows for more efficient processing and reduces the risk of timeout errors. Additionally, using SQL activities to merge and process the data in parallel can help to reduce the overall processing time.

For example, if the data has 80+ attributes, it may be possible to split the data into 4-6 tables, each with a subset of the attributes. The SQL activity can then be used to join the tables and process the data in parallel.

Example Code

SQL Activity Example

SELECT * FROM Table1
JOIN Table2 ON Table1.ID = Table2.ID
JOIN Table3 ON Table2.ID = Table3.ID

Best Practices

Checklist for Optimizing Data Extension Imports

  • Split large datasets into smaller, normalized tables
  • Use SQL activities to merge and process the data in parallel
  • Optimize the import process by reducing the number of attributes and using efficient data types
  • Use Automation Studio to schedule and manage the import process
  • Monitor the import process and adjust as needed to avoid timeout errors
  • Consider using external data sources, such as Salesforce Data Cloud, to offload large datasets

Frequently Asked Questions

What is the maximum number of records that can be imported into a data extension?

The maximum number of records that can be imported into a data extension is not strictly limited, but loading 100 million records into a single data extension is considered an extreme use case and can result in performance issues.

How can I optimize the import process for large datasets?

To optimize the import process, split the data into smaller, normalized tables, and use SQL activities to merge and process the data in parallel. Additionally, optimize the import process by reducing the number of attributes and using efficient data types.

What are the benefits of using Automation Studio to manage the import process?

Using Automation Studio to manage the import process allows for scheduling and automation of the import process, reducing the risk of human error and improving overall efficiency.

Can I use external data sources, such as Salesforce Data Cloud, to offload large datasets?

Yes, using external data sources, such as Salesforce Data Cloud, can help to offload large datasets and improve overall performance.

How can I monitor the import process and avoid timeout errors?

To monitor the import process and avoid timeout errors, use Automation Studio to schedule and manage the import process, and adjust as needed to avoid timeout errors.

Blogs for the

Business-Savvy!​

Let’s Connect

A 30 min no cost strategy session
with cloud support expert

Let’s Connect

A 30 min no cost strategy session
with cloud support expert