Document Type: Framework
Status: Active
Authority: Data Brain
Parent: Data Brain Architecture
Applies To: All MWMS environments where behavioural, performance, or measurement signals are used for analysis, optimisation, or decision-making
Version: v1.0
Last Reviewed: 2026-04-23
Purpose
The Data Brain Visibility Gap Framework defines how MWMS identifies, classifies, and accounts for areas where data cannot be observed, collected, or trusted.
Not all user behaviour is visible.
Not all systems allow tracking.
Not all interactions can be measured.
This framework ensures that:
• missing data is not misinterpreted
• unobservable behaviour is acknowledged
• decisions account for blind spots
• system confidence reflects real visibility limits
Without visibility awareness, MWMS risks treating incomplete data as complete truth.
Core Principle
Absence of data does not mean absence of behaviour.
A missing signal may represent:
• untracked behaviour
• blocked tracking
• inaccessible environments
• technical limitations
Not all gaps can be eliminated.
Some gaps must be understood and managed.
Position in MWMS System
This framework operates within:
• Data Brain → measurement awareness
• Research Brain → insight interpretation
• Ads Brain → performance analysis
• Experimentation Brain → test interpretation
• HeadOffice → decision control
This framework feeds:
• Data Brain Data Trust Framework
• Data Brain Attribution Reliability Framework
• HeadOffice Data Decision Gate Framework
What Is a Visibility Gap
A visibility gap is any situation where:
• behaviour occurs but is not captured
• data is partially captured
• data is inaccessible
• tracking is technically impossible or restricted
Visibility Gap Categories
1. Technical Visibility Gaps
Caused by platform or browser limitations.
Examples:
• page unload before tracking completes
• blocked scripts
• ad blockers
• browser privacy restrictions
2. Embedded Environment Gaps
Caused by restricted access to embedded systems.
Examples:
• iframes
• third-party forms
• payment processors
• external booking systems
• chat widgets
In these environments:
• DOM access may be restricted
• events may not be exposed
• tracking containers may not be installable
3. External System Gaps
Caused by third-party systems outside control.
Examples:
• SaaS platforms
• vendor tools
• affiliate networks
• external analytics systems
These systems may:
• not expose event data
• provide incomplete data
• use different measurement logic
4. Attribution Visibility Gaps
Caused by incomplete cross-channel visibility.
Examples:
• cross-device behaviour
• offline conversions
• delayed conversions
• multi-touch interactions not fully visible
Attribution will always have partial visibility.
5. Context Visibility Gaps
Caused by missing contextual information.
Examples:
• click without region context
• interaction without funnel stage
• behaviour without user state
The signal exists, but meaning is incomplete.
6. Data Loss Gaps
Caused by tracking failure or system issues.
Examples:
• events not firing
• missing data layer values
• broken tags
• deployment errors
These gaps may be temporary or persistent.
Visibility Gap Identification
MWMS must actively identify gaps through:
• audits
• validation protocols
• anomaly detection
• cross-platform comparison
• manual testing
Gaps must not be assumed — they must be discovered.
Visibility Gap Awareness Rule
All data interpretation must consider:
• what is visible
• what is partially visible
• what is not visible
Ignoring visibility gaps leads to false conclusions.
Visibility Gap Impact Levels
High Impact Gap
Conditions:
• critical behaviour not visible
• major part of funnel missing
• key conversion steps untracked
Impact:
→ decisions may be unsafe
Medium Impact Gap
Conditions:
• partial visibility
• context missing
• attribution incomplete
Impact:
→ decisions require caution
Low Impact Gap
Conditions:
• minor or non-critical gaps
• limited effect on interpretation
Impact:
→ minimal risk
Visibility Gap Handling Strategies
1. Acknowledge the Gap
Do not ignore or hide missing data.
Explicitly document:
• where the gap exists
• what is missing
• potential impact
2. Adjust Interpretation
Interpret data with awareness of limitations.
Example:
• low conversions may reflect missing tracking, not low performance
3. Use Supporting Signals
Where direct tracking is not possible:
• use proxy metrics
• use behavioural patterns
• use multi-signal validation
4. Reduce the Gap Where Possible
Where feasible:
• improve tracking implementation
• add data layer enhancements
• integrate systems
• enable event exposure
5. Accept Irreducible Gaps
Some gaps cannot be removed.
Examples:
• cross-device behaviour
• third-party black-box systems
These must be:
→ accepted and accounted for
Visibility Gap and Data Trust
Data trust must be adjusted based on:
• presence of visibility gaps
• severity of gaps
• ability to validate signals
High visibility gaps reduce trust.
Visibility Gap and Attribution
Attribution must consider:
• incomplete journey visibility
• hidden touchpoints
• untracked interactions
Attribution confidence decreases when gaps increase.
Visibility Gap and Experimentation
Experiments must consider:
• missing data affecting results
• incomplete funnel visibility
• distorted conversion signals
Confidence must be adjusted accordingly.
🔴 Visibility Gap Misinterpretation Risk
Common mistake:
→ assuming missing data = no behaviour
Correct interpretation:
→ missing data = unknown behaviour
🔴 Decision Risk Rule
Decisions must be adjusted or blocked when:
• high-impact visibility gaps exist
• key signals are missing
• interpretation cannot be validated
Relationship to Other Frameworks
Supports:
• Data Brain Event Reliability Framework
• Data Brain Signal Context Framework
• Data Brain Measurement Integrity Framework
• Data Brain Data Trust Framework
• Data Brain Attribution Reliability Framework
• HeadOffice Data Decision Gate Framework
Failure Modes Prevented
false conclusions from missing data
overconfidence in incomplete measurement
incorrect attribution interpretation
misleading funnel analysis
scaling decisions based on partial visibility
Drift Protection
The system must prevent:
• visibility gaps emerging unnoticed
• system changes increasing blind spots
• reliance on outdated tracking assumptions
• ignoring new limitations introduced by platforms
Architectural Intent
The Data Brain Visibility Gap Framework ensures MWMS operates with:
→ honest awareness of what it does NOT know
This is critical for:
• decision integrity
• system credibility
• long-term optimisation accuracy
Final Rule
If visibility is incomplete:
→ conclusions must be treated as partial
Change Log
Version: v1.0
Date: 2026-04-23
Author: Data Brain
Change:
Initial creation of Data Brain Visibility Gap Framework defining how MWMS identifies and manages unobservable data conditions.
Change Impact Declaration
Pages Created:
Data Brain Visibility Gap Framework
Pages Updated:
None
Pages Deprecated:
None
Registries Requiring Update:
MWMS Architecture Registry
Canon Version Update Required:
No
Change Log Entry Required:
Yes