ePrivacy and GPDR Cookie Consent by Cookie Consent Skip to main content

Step By Step Implementation Guide

Data Sources Implementation 

Get credentials

  • Request for access or find credentials to data source in 1Password vault dedicated for the client
  • Request for URL and password to the monitoring account of client's instance from #help_it
  • Confirm that the credentials are valid ASAP after received
  • Confirm access to data sources match with the defined scope of work (that we are given accesses / credentials to all the data sources in scope and none of them are out of scope) 
  • Request for sample of data (Whether the access is available or not) - this should give the analyst a good overview of what to expect 

Initial data exploration

  • Prepare data sample description following the Data Exploration checklist; paying attention to:  
    1. List of fields
    2. How sparse the data is
    3. Data values sample (SELECT DISTINCT something)
    4. Distribution
    5. Volume
  • Prepare customer events description which includes the following. Also refer to How to Format Events 
    1. Types of events (eg: purchase, shipment, delivery, cancellation)
    2. Example of payload
  • Prepare a list of unique IDs from all events
    • Compare this against the identity graph defined with the client. For each data source's events, we should see the same types of unique IDs
    • Check if there is any kind of data validation necessary article on unique ids data validation? 
  • Check the format of PS identifiers across all data sources:
    • "None", "0", "NULL" in text format can result in stitching huge customer entities
    • Is the format of the identifier the same across all data sources? "12345" and "12345.0" are different values for PS algorithm
    • Are there placeholder values which should be removed? 

Data source implementation

Implement the workspace:

  1. Extract
  2. Clean (exclude events we don’t need, remove unnecessary fields to remove, clean the values)
  3. Validate the Profile Stitching IDs
  4. Confirm the valid version/branch of format_events and cdp_db_loader components
  5. Write in Documentation:
    • General description, business assumptions of the workspace 
    • Comments in configs

Please refer to the following documentations:

Review, Finalize

  • Code review - Assign a senior analyst to do code review
  • Final checks:
    1. Double-check through queries
    2. Benchmarks with source data (eg. validate total web visits against GA numbers or total email subscribers with source. This is a quick way to identify if there are significant gaps in our data ingestion. 

Profile Stitching

Define profile stitching configuration 
Run PS workspace

Check the quality of Profile Stitching after running. 

In each project, there is always the potential to have edge cases where we wrongly stitched some customer profiles or create super entities. Refer to this guide for cleaning wrongly stitched customer entities without re-running PS, which can be a time and resource-consuming process. 

If the situation calls for it you can re-define the PS configuration and re-run profile stitching from scratch.


In the business requirement gathering process, the PM should have defined a list of required attributes from the client following this Attributes Library template. 

Based on this list, analysts should first verify all the attributes requested that the data or events for them are available and possible to calculate. Double-check on the formula we should be using as well. 

Set first & last name as the first two attributes in the customer profile
Implement SQL queries for calculation & run CPS: Attributes Calculation configuration


Data Destination

  • Before configuring data destinations, ensure that you have the relevant access/credentials to the destination in 1Password
  • OAuth repo
    1. Prepare OAuth repo, if relevant
    2. Confirm that OAuth repo is ready and the loader component can be authorized
  • Implementation
    1. Set up the connection between MI and CDP (create admin “users” for API connection, test)
    2. Check what are the requirements for the data in the destination - depends on the destination, certain IDs need to be set as exported attributes
    3. Prepare exported attributes
    4. Set export destination in MI and MBE
    5. Set up alerts for: MI Workspace and Meiro Events monitoring dashboard. This is crucial for making sure that we are always aware of failed workspaces and issues on implementation before the impact is noticed by the end user of the CDP. 
    6. Test segment export with users - send a non-empty segment to the customer, confirm that the end destination receives the segment export.

Final CDP and BE Settings

  • Tags
  • Insights
  • Colors
  • Event snippets
  • Smart segment
  • Widgets
  • User roles

Project Managers need to set up the Business Explorer instance

Welcome Emails & Product Newsletters


Handover & Documentation