Stages 03–06
Normalise. Prep.
Govern.
Four tightly coupled stages that take your structurally mapped source data and produce clean, governed, semantically modelled golden records — ready for target mapping and ERP load.
Structural Normalise
The step data engineers skip. Xynage detects type mismatches across columns, composite primary keys, orphaned columns with no downstream consumers, and naming convention conflicts. The agent proposes a normalised DDL. Your engineer reviews and approves before any data moves.
Input
Schema report + FK graph from Schema Detective
Output
Approved normalised DDL, structural diff, migration script
Data Prep
LLM-assisted cleaning rules — format standardisation, null strategies, value normalisation (country codes, phone numbers, date formats), PII detection and masking before any LLM call. Every cleaning rule is proposed by the agent and approved by your team. Cleaning log is committed to git.
Input
Normalised staging tables
Output
Clean prep tables, approved cleaning rules, PII audit log
Semantic Model
The agent generates a dbt semantic model from the clean prep tables — dimension hierarchies, measure definitions, entity relationships. You approve the model before it deploys. The result is a reproducible mart layer that downstream BI tools and ML pipelines can consume directly.
Input
Clean prep tables
Output
dbt semantic model, mart definitions, metric layer
MDM — Master Data Management
Probabilistic entity resolution using Splink with configurable blocking rules and EM training. The agent surfaces duplicate candidate pairs ranked by match score. Stewards review and approve or reject each merge decision. Every decision is logged to the approver. Golden records are versioned — merge history preserved.
Input
Clean prep tables
Output
Golden records, merge decision log, survivorship rules, versioned merge history
Previous
← Schema DetectiveNext stage
Target Mapping + Activate →