This section dives into how data enters Salesforce Data Cloud and how it is structured for use. Understanding data ingestion and modeling is key for managing and leveraging customer data effectively.
Data ingestion refers to the process of bringing data into Salesforce Data Cloud. There are two main methods for ingesting data: batch ingestion and real-time ingestion.
Batch ingestion involves importing large volumes of data at once. This is suitable for historical data or scheduled updates.
Key Details:
Example:
A company wants to upload all sales data from the past year. They export this data as a CSV file from their ERP system and import it into Data Cloud using a batch upload tool.
Real-time ingestion is designed to capture and process data as it happens, making it ideal for time-sensitive actions.
Key Details:
Example:
A retail website captures every product click in real time. If a customer views a product but doesn’t purchase it, the system immediately sends a personalized follow-up email.
Data connectors are tools that allow Salesforce Data Cloud to integrate with various external systems and bring in data from multiple sources.
Example:
A marketing team links Google Ads to Salesforce Data Cloud to automatically import campaign performance data every day.
Once data is ingested, it must be organized into a structure that allows easy access and analysis. This process is called data modeling.
Data objects are like tables in a database, representing key entities such as customers, transactions, or products.
Key Points:
Relationships link different objects to reflect their connections. For example:
How to Design Relationships:
Example:
Sarah (a customer) places two orders. Her Customer ID links her to both orders. Each order includes products linked through a Product ID.
Data extensions allow you to customize the standard data model to meet specific business needs.
When to Use Data Extensions:
Example:
A travel company might add custom fields to track "Preferred Destinations" and "Frequent Flyer Miles" for customers.
Before using data, it must be cleaned to ensure accuracy and consistency.
Why Data Cleaning Matters:
Data Ingestion Processes:
Data Modeling Principles:
Use Cases for Ingestion Methods:
Data ingestion and modeling are foundational to Salesforce Data Cloud:
Mastering these processes will help you ensure a clean, structured, and actionable dataset in Salesforce Data Cloud.
When ingesting external data into Salesforce Data Cloud, different systems often have incompatible structures and formats. Proper Data Mapping ensures that data is correctly aligned with the Data Cloud schema, preventing ingestion errors and maintaining data integrity.
Field Mapping:
Data Transformation:
MM/DD/YYYY → YYYY-MM-DD$1,000 → 1000.00Data Normalization:
+1 555-1234 across all sources.| External Data Field | Salesforce Data Cloud Field |
|---|---|
user_email |
Customer Email |
purchase_date |
Order Date |
total_spent |
Order Amount |
Poor data quality due to errors or missing values can lead to faulty analytics, incorrect insights, and compliance risks. Salesforce Data Cloud applies Data Validation Rules to ensure that incoming data meets accuracy, completeness, and consistency requirements.
Customer ID must always be provided.Order Date cannot be later than Current Date.| Incorrect Data | Validation Rule | Corrected Data |
|---|---|---|
john.doe@email |
Email must contain @domain.com |
[email protected] |
2025-13-01 |
Date format invalid | 2025-12-01 |
Order Amount = -100 |
Order amount cannot be negative | Order Amount = 100 |
As businesses process large volumes of data, optimizing data ingestion is essential to reduce processing time, minimize system load, and ensure near real-time updates.
.zip, .gz).A retail business processes 1 million orders daily:
| Optimization Step | Processing Time Before | Processing Time After |
|---|---|---|
| Full dataset refresh | 3 hours | N/A |
| Incremental updates (new orders only) | N/A | 15 minutes |
To comply with GDPR, CCPA, and other privacy regulations, companies must ensure that customer data is handled securely and only accessible to authorized users.
| Regulation | Requirement | Data Cloud Solution |
|---|---|---|
| GDPR | Customers can request deletion of their data | "Delete Request API" processes requests automatically |
| CCPA | Customers can opt out of data sharing | Data Cloud ensures opt-out preferences are honored |
By mastering these concepts, businesses can efficiently ingest and model high-quality, compliant data in Salesforce Data Cloud.
What is the difference between a Data Lake Object (DLO) and a Data Model Object (DMO) in Salesforce Data Cloud?
A Data Lake Object stores raw ingested data, while a Data Model Object represents standardized customer data used for identity resolution, segmentation, and analytics.
When data is ingested into Data Cloud through a data stream, it first lands in a Data Lake Object (DLO). This object preserves the source structure and stores the raw dataset exactly as it arrives from the external system.
After ingestion, the data is mapped to Data Model Objects (DMOs). DMOs follow Salesforce’s Customer 360 data model, which standardizes attributes such as individuals, contact points, orders, and engagement events.
This transformation layer allows Data Cloud to unify data from many sources that may have different schemas.
A common mistake is trying to run segmentation directly on DLOs. Instead, segmentation and identity resolution operate on DMOs, not raw lake data.
Demand Score: 88
Exam Relevance Score: 90
How does a Data Stream move data from a source system into Salesforce Data Cloud?
A data stream connects to an external data source, ingests the data into Data Lake Objects, and then maps the data into Data Model Objects.
Data streams act as the ingestion pipelines of Data Cloud. They define how data is collected from external systems such as Salesforce CRM, Amazon S3, Snowflake, or web engagement platforms.
The ingestion process occurs in two major steps:
Extraction – Data is pulled from the source system.
Landing – The data is stored in Data Lake Objects.
After landing, administrators configure mapping rules to map the source fields into the standardized Data Model Objects.
This mapping step is critical because it allows identity resolution and segmentation to operate on normalized customer data across multiple systems.
Demand Score: 85
Exam Relevance Score: 92
When should you create a custom Data Model Object instead of using standard DMOs?
A custom Data Model Object should be created when the data does not fit into the standard Customer 360 schema provided by Data Cloud.
Salesforce provides many standard DMOs such as Individual, Contact Point Email, Engagement, and Order. These cover common customer data use cases.
However, organizations sometimes ingest data that does not match these predefined objects. Examples include proprietary loyalty systems, custom subscription models, or industry-specific data like insurance policies.
In these cases, administrators create a custom DMO to store the data while still integrating it with the broader customer profile architecture.
Even when custom objects are created, they should still include identifiers that allow them to connect to the Individual DMO, ensuring the data can participate in identity resolution and segmentation.
Demand Score: 81
Exam Relevance Score: 86
Why is field mapping important when ingesting data into Data Cloud?
Field mapping ensures that source data attributes are correctly aligned with the standardized fields in Data Model Objects.
Different systems often store the same information using different field names or formats. For example, a marketing system may store an email address as emailAddress, while a CRM system uses Email.
Field mapping resolves these differences by assigning each source field to the appropriate DMO attribute.
Correct mapping is essential because identity resolution, segmentation, and calculated insights rely on consistent attribute definitions.
Incorrect mappings can lead to identity matching failures, incomplete customer profiles, or inaccurate segmentation results.
Demand Score: 84
Exam Relevance Score: 87
What types of data sources can be ingested into Salesforce Data Cloud?
Data Cloud can ingest data from Salesforce systems, data warehouses, cloud storage platforms, and streaming event sources.
The platform supports multiple ingestion mechanisms so organizations can unify data from many environments.
Common data sources include:
Salesforce CRM objects
Data warehouses like Snowflake or BigQuery
Cloud storage such as Amazon S3
Web and mobile engagement events
Marketing and commerce systems
Data is typically ingested through connectors or batch file imports.
Supporting multiple ingestion types allows Data Cloud to build a complete customer profile across operational, behavioral, and transactional data sources.
Demand Score: 80
Exam Relevance Score: 82
What is the role of data transformations in the Data Cloud ingestion process?
Data transformations clean, standardize, or reshape ingested data before it is used for identity resolution or analytics.
Source data often contains inconsistencies such as different date formats, missing fields, or inconsistent naming conventions.
Data Cloud transformations allow administrators to:
Normalize attribute formats
Combine or split fields
Apply calculated values
Standardize identifiers
These transformations ensure that data conforms to the Customer 360 data model before downstream processes like identity resolution or segmentation occur.
Without transformation steps, inconsistent data structures could prevent records from matching correctly or cause inaccurate analytics results.
Demand Score: 83
Exam Relevance Score: 86