Data Subject Request API Version 1 and 2
Data Subject Request API Version 3
Platform API Overview
Accounts
Apps
Audiences
Calculated Attributes
Data Points
Feeds
Field Transformations
Services
Users
Workspaces
Warehouse Sync API Overview
Warehouse Sync API Tutorial
Warehouse Sync API Reference
Data Mapping
Warehouse Sync SQL Reference
Warehouse Sync Troubleshooting Guide
ComposeID
Warehouse Sync API v2 Migration
Bulk Profile Deletion API Reference
Calculated Attributes Seeding API
Custom Access Roles API
Data Planning API
Group Identity API Reference
Pixel Service
Profile API
Events API
mParticle JSON Schema Reference
IDSync
AMP SDK
Initialization
Configuration
Network Security Configuration
Event Tracking
User Attributes
IDSync
Screen Events
Commerce Events
Location Tracking
Media
Kits
Application State and Session Management
Data Privacy Controls
Error Tracking
Opt Out
Push Notifications
WebView Integration
Logger
Preventing Blocked HTTP Traffic with CNAME
Linting Data Plans
Troubleshooting the Android SDK
API Reference
Upgrade to Version 5
Direct URL Routing FAQ
Web
Android
iOS
Cordova Plugin
Identity
Initialization
Configuration
Event Tracking
User Attributes
IDSync
Screen Tracking
Commerce Events
Location Tracking
Media
Kits
Application State and Session Management
Data Privacy Controls
Error Tracking
Opt Out
Push Notifications
Webview Integration
Upload Frequency
App Extensions
Preventing Blocked HTTP Traffic with CNAME
Linting Data Plans
Troubleshooting iOS SDK
Social Networks
iOS 14 Guide
iOS 15 FAQ
iOS 16 FAQ
iOS 17 FAQ
iOS 18 FAQ
API Reference
Upgrade to Version 7
Getting Started
Identity
Upload Frequency
Getting Started
Opt Out
Initialize the SDK
Event Tracking
Commerce Tracking
Error Tracking
Screen Tracking
Identity
Location Tracking
Session Management
Getting Started
Identity
Initialization
Configuration
Content Security Policy
Event Tracking
User Attributes
IDSync
Page View Tracking
Commerce Events
Location Tracking
Media
Kits
Application State and Session Management
Data Privacy Controls
Error Tracking
Opt Out
Custom Logger
Persistence
Native Web Views
Self-Hosting
Multiple Instances
Web SDK via Google Tag Manager
Preventing Blocked HTTP Traffic with CNAME
Facebook Instant Articles
Troubleshooting the Web SDK
Browser Compatibility
Linting Data Plans
API Reference
Upgrade to Version 2 of the SDK
Web
Alexa
Overview
Step 1. Create an input
Step 2. Verify your input
Step 3. Set up your output
Step 4. Create a connection
Step 5. Verify your connection
Step 6. Track events
Step 7. Track user data
Step 8. Create a data plan
Step 9. Test your local app
Overview
Step 1. Create an input
Step 2. Verify your input
Step 3. Set up your output
Step 4. Create a connection
Step 5. Verify your connection
Step 6. Track events
Step 7. Track user data
Step 8. Create a data plan
Step 1. Create an input
Step 2. Create an output
Step 3. Verify output
Node SDK
Go SDK
Python SDK
Ruby SDK
Java SDK
Introduction
Outbound Integrations
Firehose Java SDK
Inbound Integrations
Compose ID
Data Hosting Locations
Glossary
Migrate from Segment to mParticle
Migrate from Segment to Client-side mParticle
Migrate from Segment to Server-side mParticle
Segment-to-mParticle Migration Reference
Rules Developer Guide
API Credential Management
The Developer's Guided Journey to mParticle
Create an Input
Start capturing data
Connect an Event Output
Create an Audience
Connect an Audience Output
Transform and Enhance Your Data
The new mParticle Experience
The Overview Map
Introduction
Data Retention
Connections
Activity
Live Stream
Data Filter
Rules
Tiered Events
mParticle Users and Roles
Analytics Free Trial
Troubleshooting mParticle
Usage metering for value-based pricing (VBP)
Introduction
Sync and Activate Analytics User Segments in mParticle
User Segment Activation
Welcome Page Announcements
Project Settings
Roles and Teammates
Organization Settings
Global Project Filters
Portfolio Analytics
Analytics Data Manager Overview
Events
Event Properties
User Properties
Revenue Mapping
Export Data
UTM Guide
Data Dictionary
Query Builder Overview
Modify Filters With And/Or Clauses
Query-time Sampling
Query Notes
Filter Where Clauses
Event vs. User Properties
Group By Clauses
Annotations
Cross-tool Compatibility
Apply All for Filter Where Clauses
Date Range and Time Settings Overview
Understanding the Screen View Event
Analyses Introduction
Getting Started
Visualization Options
For Clauses
Date Range and Time Settings
Calculator
Numerical Settings
Assisted Analysis
Properties Explorer
Frequency in Segmentation
Trends in Segmentation
Did [not] Perform Clauses
Cumulative vs. Non-Cumulative Analysis in Segmentation
Total Count of vs. Users Who Performed
Save Your Segmentation Analysis
Export Results in Segmentation
Explore Users from Segmentation
Getting Started with Funnels
Group By Settings
Conversion Window
Tracking Properties
Date Range and Time Settings
Visualization Options
Interpreting a Funnel Analysis
Group By
Filters
Conversion over Time
Conversion Order
Trends
Funnel Direction
Multi-path Funnels
Analyze as Cohort from Funnel
Save a Funnel Analysis
Explore Users from a Funnel
Export Results from a Funnel
Saved Analyses
Manage Analyses in Dashboards
Dashboards––Getting Started
Manage Dashboards
Organize Dashboards
Dashboard Filters
Scheduled Reports
Favorites
Time and Interval Settings in Dashboards
Query Notes in Dashboards
User Aliasing
The Demo Environment
Keyboard Shortcuts
Analytics for Marketers
Analytics for Product Managers
Compare Conversion Across Acquisition Sources
Analyze Product Feature Usage
Identify Points of User Friction
Time-based Subscription Analysis
Dashboard Tips and Tricks
Understand Product Stickiness
Optimize User Flow with A/B Testing
User Segments
IDSync Overview
Use Cases for IDSync
Components of IDSync
Store and Organize User Data
Identify Users
Default IDSync Configuration
Profile Conversion Strategy
Profile Link Strategy
Profile Isolation Strategy
Best Match Strategy
Aliasing
Overview
Create and Manage Group Definitions
Introduction
Catalog
Live Stream
Data Plans
Blocked Data Backfill Guide
Predictive Audiences Overview
Using Predictive Audiences
Predictive Attributes Overview
Create Predictive Attributes
Assess and Troubleshoot Predictions
Use Predictive Attributes in Campaigns
Introduction
Profiles
Warehouse Sync
Data Privacy Controls
Data Subject Requests
Default Service Limits
Feeds
Cross-Account Audience Sharing
Approved Sub-Processors
Import Data with CSV Files
CSV File Reference
Glossary
Video Index
Single Sign-On (SSO)
Setup Examples
Introduction
Introduction
Introduction
Rudderstack
Google Tag Manager
Segment
Advanced Data Warehouse Settings
AWS Kinesis (Snowplow)
AWS Redshift (Define Your Own Schema)
AWS S3 Integration (Define Your Own Schema)
AWS S3 (Snowplow Schema)
BigQuery (Snowplow Schema)
BigQuery Firebase Schema
BigQuery (Define Your Own Schema)
GCP BigQuery Export
Snowplow Schema Overview
Snowflake (Snowplow Schema)
Snowflake (Define Your Own Schema)
Aliasing
Event
Audience
Event
Audience
Feed
Event
Audience
Cookie Sync
Event
Audience
Audience
Audience
Feed
Event
Event
Event
Event
Audience
Event
Data Warehouse
Event
Event
Event
Audience
Event
Feed
Event
Event
Event
Event
Audience
Event
Event
Event
Feed
Event
Event
Audience
Feed
Event
Event
Custom Feed
Event
Data Warehouse
Event
Event
Audience
Audience
Audience
Event
Audience
Event
Event
Event
Event
Event
Audience
Audience
Event
Event
Audience
Data Warehouse
Event
Cookie Sync
Audience
Event
Event
Event
Event
Event
Feed
Feed
Event
Event
Event
Audience
Event
Event
Audience
Event
Event
Event
Feed
Audience
Event
Audience
Event
Audience
Event
Audience
Audience
Audience
Audience
Event
Event
Event
Event
Event
Feed
Event
Event
Event
Event
Event
Feed
Audience
Event
Event
Event
Event
Event
Event
Feed
Event
Audience
Event
Event
Event
Custom Pixel
Feed
Event
Event
Event
Audience
Event
Event
Event
Data Warehouse
Event
Event
Audience
Audience
Audience
Event
Audience
Audience
Audience
Cookie Sync
Event
Feed
Audience
Event
Event
Audience
Audience
Event
Event
Event
Event
Audience
Cookie Sync
Audience
Cookie Sync
Feed
Audience
Event
mParticle’s Data Warehouse integration with Snowflake forwards all your incoming data to a Snowflake cluster, allowing you to query the raw data directly.
The integration creates a table in your Snowflake database for each custom event name and each eCommerce event name with a volume above a defined threshold. Less common events are recorded in a single table, labeled otherevents
.
By default, the integration begins loading current data into Snowflake from the time it is enabled. You can work with your mParticle Customer Service Manager to load historical data.
All setup tasks can be accomplished from a Snowflake Worksheet.
Note that you can use any names you choose for your warehouse, database, schema, role, and user, as long as you provide the correct names to mParticle in the integration settings. Also note that we currently don’t support Snowflake’s double-quoted identifiers (https://docs.snowflake.net/manuals/sql-reference/identifiers-syntax.html#double-quoted-identifiers), and thus please make sure no double quotes are used when you create your warehouse, database, schema, role, and user.
-- Create a warehouse and choose the appropriate size. We use AUTO_SUSPEND of 10 minutes (600 seconds) as an example. Please adjust accordingly if needed.
CREATE WAREHOUSE mPTravelWarehouse WITH WAREHOUSE_SIZE = 'XSMALL' AUTO_SUSPEND = 600 AUTO_RESUME = TRUE;
-- Create database
CREATE DATABASE mPTravelDatabase;
-- Create schema
CREATE SCHEMA mPTravelSchema WITH managed access;
Once your database is ready, you need to create a dedicated role with permissions to manage the database.
-- Create new role:
CREATE ROLE data_loader;
-- Grant access to your warehouse, database and schema
GRANT USAGE ON WAREHOUSE mPTravelWarehouse TO ROLE data_loader;
GRANT USAGE ON DATABASE mPTravelDatabase TO ROLE data_loader;
GRANT ALL ON SCHEMA mPTravelSchema TO ROLE data_loader;
-- Create user with your new role. Make sure to set your new user as a LEGACY_SERVICE user to avoid Snowflake's MFA requirement.
CREATE USER mparticle_user
MUST_CHANGE_PASSWORD = FALSE
DEFAULT_ROLE = data_loader
PASSWORD = "STRONG_PASSWORD_HERE"
TYPE = LEGACY_SERVICE;
GRANT ROLE data_loader TO USER mparticle_user;
After adding Snowflake from the integrations Directory, you can find the settings UI at Setup > Data Warehouse.
From the main page for your Snowflake configuration, select the Settings tab to provide the necessary settings to get your Snowflake integration working.
Your full account name may include region. For example, xy12345.us-east-1
.
To forward data subject erasure requests to Snowflake, set the Forwarding Status toggle to Active and select I understand after reading the disclaimer. Once the status has been set to Active, erasure requests are sent to Snowflake immediately upon being received by mParticle.
Setting Name | Data Type | Default Value | Description |
---|---|---|---|
Account | string |
Your Snowflake account name. Your full account name may include region. | |
Database Name | string |
The database name created in your Snowflake setup. | |
Data Warehouse Name | string |
The warehouse name created in your Snowflake setup. | |
User ID | string |
User ID for the user you created in your Snowflake setup. These credentials will be used to manage the schema and load data. | |
User Password | string |
The password for the user created in your Snowflake setup. | |
Schema Name | string |
The name of the schema created in your Snowflake setup. | |
Events Threshold | number |
10000 | The number of times a custom or commerce event name must be received in a 30 day period for mParticle to create a dedicated table for that event. |
Single Table Mode | boolean |
false | If enabled, events will be saved in a single table and the Events Threshold setting will not be applied. |
Configuration Name | string |
The name you are giving to this configuration. | |
Delay Between Loading Sessions in Minutes | number |
15 | Allows you to adjust how often you want to load data into the data warehouse. Note that the minimum time is 1 minute and the maximum time is 24 hours (60 minutes x 24). |
If you check the Use same settings for Development and Production box, the same configuration is used for both development and production environments.
Once your Data Warehouse integration is configured, connect individual inputs to the Snowflake output from the Connections page. You must connect every input for which you want to store data.
Setting Name | Data Type | Default Value | Platform | Description |
---|---|---|---|---|
Snowflake Table Name | string |
Feed | Table name for this partner feed. If not set, the partner name will be used. Only applicable to feeds inputs, no effect on apps inputs. If “Split Partner Feed Data by Event Name” checkbox is enabled, this setting is not used. | |
Split Partner Feed Data by Event Name | boolean |
False | Feed | If enabled, split partner feed data by event name. Otherwise load data into the same table. |
Send Batches without Events | boolean |
True | All | If enabled, an event batch that contains no events will be forwarded. |
All tables created in Snowflake have the same schema, consisting of a single column of type variant
(a dedicated Snowflake type to efficiently handle JSON data) with the name "data"
. Each row in a table is a JSON string with multiple key/value pairs.
For example:
{
"accumulatedltvvalue": 0,
"accuracy": 2,
"appenvironment": "Development",
"appid": 4245,
"applicationbuildnumber": "2",
"appname": "Acme testing",
"appplatformid": 8140,
"appversion": "2.0",
"audiencemembership": ["123", "456", "789"],
"batchid": -6520741417792989986,
"batchtimestamp": 1553009917073,
"brand": "google",
"cityname": "Sierra View",
"clientip": "75.154.15.95",
"clientipv6": "75.154.15.95",
"countrycode": "US",
"customerid": "9172349@gmail.com",
"dataconnectiontype": "wifi",
"devicemodel": "Nexus 7",
"devicename": "Unknown",
"deviceutcoffset": -5,
"email": "7309226@acme.com",
"entrypointtype": 128,
"eventattributes": {
"$Amount": "5.37769004487325",
"Navigation 0 Attr 0": "12.3",
"another new attribute": "value",
"first_name": "First",
"last_name": "Last",
"newattribute": "value",
"yet another new attribute": "value"
},
"eventdate": "2019-03-19",
"eventhour": "2019-03-19 15:00:00",
"eventid": 351939524822094163,
"eventlength": 0,
"eventltvvalue": 5.37769004487325,
"eventname": "Navigation 0",
"eventstarttimestamp": 1553009774253,
"eventtimestamp": 1553009774253,
"eventtypeid": 1,
"firstseentimestamp": 1553009917073,
"googleaid": "9a8cd090-1a4f-4cb9-b76f-1bbca598d985",
"isdebug": true,
"latitude": 41.033192,
"localecountry": "US",
"localelanguage": "EN",
"longitude": -75.449047,
"manufacture": "LGE",
"messagetypeid": 4,
"mparticleuserid": -5045766802590845105,
"networkcarrier": "Sprint",
"networkcountry": "US",
"osversion": "4.2.1",
"osversionint": 0,
"packagename": "com.mparticle.demo",
"platform": "Android",
"product": "occam",
"regioncode": "PA",
"screendpi": 160,
"screenheight": 736,
"screenwidth": 1280,
"sdkversion": "5.1.0",
"sessionid": 3021067757087817833,
"sessionstarttimestamp": 1553009774253,
"upgradedate": 0,
"userattributes": {
"$Age": "85",
"$Gender": "male",
"$Zip": "95450",
"LiveInNewYork": "true",
"another new user attribute": "56",
"status": "gold"
},
"workspaceid": 4254,
"yahoouserid": "1940141@yahoo.com"
}
mParticle also creates two types of views under the schema:
mp_vw_{tableName}
is created that allows you to run regular SQL queries against each table. For example, to query workspaceid
from each table, instead of using data:workspaceid
to query the table, you can use workspaceid
to query the view. Each user attribute and event attribute has its own column in the view. For user attribute named Some Sample User Attribute
and event attribute named Some Sample Event Attribute
, the column name in the view is "ua Some Sample User Attribute"
and "ea Some Sample Event Attribute"
, respectively. Attribute column names have double quotes and are case sensitive.eventsview
that unions all per table views to give you easy access to all data under the schema.Use syntax data:key_name
. Here are some sample queries:
-- select some "columns" to look at, if querying the table directly
select data:appid, data:eventname, data:eventtimestamp, data:customerid, data:mparticleuserid
from sample_table
limit 10
-- select some "columns" to look at, if querying the view
select appid, eventname, eventtimestamp, customerid, mparticleuserid
from mp_vw_sample_table
limit 10
-- count unique eventid's by hour by event name
select date_trunc('hour', to_timestamp(cast(data:eventtimestamp / 1000 as int))), data:eventname, count(distinct data:eventid)
from sample_table
group by 1, 2
order by 3 desc
If you have chosen to create an IP whitelist as part of your Snowflake Network Policy, you can access a current list of IP addresses used by mParticle here.
Events from each connected Partner Feed will be stored under a single table unless the Split Partner Feed Data by Event Name checkbox is enabled. You can choose the table name for each Feed in the Connection Settings. If you do not provide a name, mParticle will use the name of the Partner.
Events can be forwarded with a Device Application Stamp stored in the device ID column. You can enable this in the settings page for your data warehouse configuration by toggling the Store Device Stamp checkbox.
mParticle loads data into Snowflake via Amazon S3 and can tolerate the Snowflake cluster being unavailable for up to 30 days, depending on data volume. In the event of extended downtime on your cluster, a data replay can be arranged.