Document Type: Protocol
Status: Active
Authority: Data Brain
Parent: Data Brain Architecture
Applies To: All tracking implementations, analytics setups, and measurement systems across MWMS
Version: v1.0
Last Reviewed: 2026-04-23
Purpose
The Data Brain Measurement Validation Protocol defines the mandatory process for verifying that measurement systems are accurate, complete, and decision-safe before data is used within MWMS.
This protocol ensures:
• events are correctly implemented
• data reflects real behaviour
• measurement errors are detected early
• invalid data is prevented from influencing decisions
This protocol acts as the approval layer for all measurement systems.
Core Principle
Data must be validated before it is trusted.
If measurement is not validated:
→ it must not be used
Position in MWMS System
This protocol operates between:
• Measurement Integrity Framework
• Data Trust Framework
• Analytics Audit Framework
It determines:
👉 whether data is valid
👉 whether data can be trusted
👉 whether data can be used for decisions
Validation Execution Flow
All measurement validation must follow this sequence:
Step 1 — Define Expected Behaviour
Before testing, define:
• expected events
• expected parameters
• expected values
• expected user journey
Example:
Landing page → CTA click → form submit → conversion event
Step 2 — Execute Test Journeys
Simulate real user behaviour:
• navigate key pages
• trigger events manually
• complete full funnel actions
Purpose:
→ confirm real-world tracking accuracy
Step 3 — Validate Event Firing
Check:
• event fires when expected
• event does not fire when not expected
Tools:
• GTM preview mode
• GA4 debug view
Step 4 — Validate Parameters
Confirm:
• parameters exist
• parameters are correctly named
• values are accurate
Example:
• correct purchase value
• correct product IDs
• correct campaign data
Step 5 — Validate Data Layer
Check:
• correct data layer structure
• correct values passed to tags
• no missing variables
Step 6 — Detect Duplicate Events
Check for:
• duplicate conversions
• duplicate page views
• multiple event firing
Indicators:
• inflated counts
• inconsistent ratios
Step 7 — Detect Missing Events
Check:
• all expected events occur
• no gaps in funnel tracking
Method:
• compare expected vs actual event list
Step 8 — Validate Funnel Continuity
Confirm:
• events occur in correct sequence
• no broken funnel steps
• no unexpected drop-offs
Step 9 — Cross-Platform Validation
Compare:
• GA4 vs Ads platform
• analytics vs backend system
Check:
• conversion consistency
• event count alignment
Step 10 — Validate Attribution Inputs
Confirm:
• UTMs captured correctly
• source/medium assigned correctly
• no unwanted referrals
Step 11 — Validate Data Stability
Check over time:
• no sudden unexplained spikes
• no sudden drops
• consistent behaviour across sessions
Validation Outcome Classification
All validation results must be classified:
Valid Measurement
Conditions:
• events correct
• no duplication
• no missing data
• stable behaviour
→ Approved for use
Conditionally Valid Measurement
Conditions:
• minor inconsistencies
• known limitations
→ Use with caution
Invalid Measurement
Conditions:
• duplicate events
• missing events
• incorrect values
• unstable behaviour
→ Must not be used
Validation Approval Rule
Data is only approved when:
• all critical events validated
• no duplication detected
• no critical gaps exist
• behaviour matches expectations
If any condition fails:
→ validation fails
Validation Triggers
Validation must be performed when:
• new tracking implemented
• GTM changes made
• new campaign launched
• funnel updated
• anomalies detected
• audits identify issues
Validation Frequency
Minimum:
• initial setup → full validation
• after changes → revalidation
• periodic → monthly checks
Common Validation Failures
This protocol detects:
• duplicate conversion tracking
• missing events
• incorrect parameter values
• broken funnel tracking
• attribution misclassification
• tracking breaks after updates
Relationship to Other Frameworks
This protocol supports:
• Data Brain Measurement Integrity Framework
• Data Brain Data Trust Framework
• Data Brain Analytics Audit Framework
• Data Brain Attribution Reliability Framework
• Experimentation Brain Statistical Confidence Framework
Key Outcomes
When applied correctly:
• measurement becomes reliable
• errors are caught early
• invalid data is blocked
• decision quality improves
• experimentation confidence increases
Change Log
Version: v1.0
Date: 2026-04-23
Author: Data Brain
Change:
Initial creation of Measurement Validation Protocol defining structured validation process for all measurement systems.
Change Impact Declaration
Pages Created:
Data Brain Measurement Validation Protocol
Pages Updated:
None
Pages Deprecated:
None
Registries Requiring Update:
MWMS Architecture Registry
Canon Version Update Required:
No
Change Log Entry Required:
Yes