Overview
A supply chain visibility product team struggled to prioritize roadmap work because third?party carrier data was inconsistent. Milestones were missing or late, identifiers drifted across feeds, and noisy events led to unreliable ETAs and alerts. Intelligex implemented a data quality scoring pipeline that validated events at ingest, produced carrier?level scorecards and dashboards, and enforced auto?suppression rules when feeds fell below agreed thresholds. Product reviews shifted to trustworthy signals, customer promises were more dependable, and PMs defended trade?offs with shared metricswithout changing integration channels or replatforming analytics.
Client Profile
- Industry: Supply chain visibility and logistics SaaS
- Company size (range): Multi?carrier, multi?modal platform serving enterprise shippers and 3PLs
- Stage: Established carrier integrations and warehouse; data quality handled ad hoc
- Department owner: Product Management & R&D
- Other stakeholders: Data Engineering/Analytics, Carrier Operations, Customer Success, Sales/RevOps, SRE/DevOps, Legal/Privacy, Support
The Challenge
Carrier feeds arrived via APIs, EDI, webhooks, and batch files. Common issues included missing departure or delivery milestones, duplicate events, timezone and clock skew, ambiguous location names, and inconsistent shipment identifiers between booking and tracking systems. Product managers struggled to decide which features to prioritizepredictive ETAs, exception alerts, or lane analyticsbecause the underlying data varied by carrier, lane, and service level. Customer?facing promises wavered when an alert looked credible one day and noisy the next.
Teams tried to patch gaps with spreadsheets and SQL snippets. Analytics reported high?level adoption, but no one could credibly answer which carriers produced reliable event chains or where timeliness broke down. Engineers added heuristics per carrier; Customer Success adjusted messaging by account; and Carrier Operations escalated issues reactively. Meetings focused on reconciling anecdotes rather than aligning on the real impact of data quality by cohort.
The organization wanted to keep its existing warehouse and pipelines while adding a governed layer for event validation and quality scoring. For standards alignment, the team referenced GS1 event models such as EPCIS. The data platform remained in Snowflake, with checks inspired by patterns from Great Expectations.
Why It Was Happening
Root causes were fragmented taxonomies and uneven validation. Carriers labeled milestones differently, locations were a mix of free text and codes, and identifiers shifted between booking, tender, and tracking systems. The ingestion path verified schema shape but not semantics such as event order or on?time windows. Without a shared definition of completeness and timeliness per carrier and mode, each team formed its own view of data quality, which led to inconsistent decisions.
Ownership was diffuse. Data Engineering focused on throughput, Carrier Operations on relationships and escalations, and Product on outcomes like ETA accuracy and alert credibility. There was no canonical quality score, no automated suppression when feeds degraded, and no dashboards tying scores to product behavior and customer commitments.
The Solution
Intelligex implemented a data quality scoring pipeline that validated events on ingest, scored carriers and lanes across dimensions, and enforced auto?suppression and fallbacks when scores dipped. A canonical event model aligned milestone names, location semantics, and identifiers across feeds. Validation rules checked completeness, timeliness, sequence order, identifier consistency, and basic geo plausibility. Scores rolled up by carrier, lane, service, and customer segment. Dashboards exposed score trends and their impact on ETAs and alerts, while suppression rules prevented unreliable events from triggering customer?facing features. Human?in?the?loop review governed thresholds and carrier exceptions.
- Integrations: Ingested APIs, EDI, webhooks, and batch files into a curated warehouse in Snowflake; transformation in the teams standard tools (for teams using dbt, see dbt docs); validation patterns aligned with Great Expectations; BI dashboards in the existing analytics surface; Jira used for carrier follow?ups and product gates.
- Canonical event model: Standardized milestone taxonomy (pickup, departure, arrival, out?for?delivery, delivered, exception), identifier mapping for booking/tender/tracking references, and location normalization with timezone handling. Optional mapping to GS1 EPCIS concepts.
- Validation checks: Rules for event order, completeness of required milestones by mode, timeliness windows, duplicate suppression, time skew correction, identifier consistency, and geo plausibility versus lane definitions.
- Quality scoring: Composite scores per carrier?lane?service combining completeness, timeliness, sequence integrity, and consistency. Effective dating preserved history across carrier changes.
- Auto?suppression and fallbacks: Routing logic muted unreliable events from driving ETAs and alerts when scores dipped; fallbacks switched to secondary signals or conservative messaging; exceptions required approval with reason codes.
- Dashboards and alerts: Carrier scorecards, lane heatmaps, and trend views; alerts to Slack/Teams when scores shifted or suppression engaged; impact panels showing affected ETAs and alert coverage.
- Product gates: Feature toggles keyed to quality thresholds (for example, proactive alerts enabled only when a carrier?lane cohort met minimum score); Jira badges showed gate status for planning.
- Governance and audit: Role?based controls for rules and thresholds, change logs, and lineage from dashboards back to validations and raw events.
Implementation
- Discovery: Cataloged carrier feeds by mode and method (API, EDI, webhook, batch); mapped common milestone names and identifier schemes; gathered examples of noisy events and high?value exceptions; reviewed SLA language and customer expectations for alerts and ETAs.
- Design: Authored the canonical event model and field mappings; defined validation checks, scoring dimensions, and thresholds; specified suppression logic and fallbacks; designed dashboards and Jira integration for carrier follow?ups and product gates; agreed on change control and approver roles.
- Build: Implemented validation jobs and scoring in the warehouse; created rollups by carrier?lane?service and customer segment; wired suppression into downstream ETA and alert pipelines; built dashboards and Teams/Slack alerts; added Jira automation for low?score carriers and product feature gates.
- Testing/QA: Ran in shadow mode to score historical and live traffic without affecting ETAs or alerts; compared results to known incidents and customer escalations; tuned thresholds, checks, and suppression rules; included a human?in?the?loop review with Product, Data, and Carrier Operations.
- Rollout: Enabled dashboards and product gates for select modes and regions first; activated suppression with conservative thresholds; expanded coverage as confidence grew; retained legacy logic as a controlled fallback during early cycles.
- Training/hand?off: Delivered sessions for PMs, Carrier Ops, Data, Support, and Sales on reading scorecards, understanding gates, and handling exceptions; updated SOPs for carrier escalations and product decisions tied to scores; transferred ownership of rules and thresholds to Product Ops and Data under change control.
Results
Roadmap and release decisions were anchored in a shared view of data quality. PMs saw which carriers and lanes produced dependable milestones and unlocked features accordingly. ETAs and alerts relied on cohorts that met thresholds, and marketing promises reflected that reality. Carrier Operations focused on a ranked list of improvements with evidence drawn from the same scorecards used in product reviews.
Operational friction eased. When a feed degraded, suppression and fallbacks prevented unreliable alerts without manual intervention, and alerts guided the right follow?ups. Support conversations referenced visible scores instead of ad hoc explanations. The warehouse, integrations, and BI tools stayed in place; the change was a governed validation and scoring layer that made signals predictable.
What Changed for the Team
- Before: Product debates leaned on anecdotes about noisy carriers. After: Carrier?level scorecards and gates framed decisions with shared metrics.
- Before: Engineers added heuristics per carrier. After: Canonical validations and suppression rules applied consistently across feeds.
- Before: Alerts and ETAs fluctuated by day. After: Features engaged only for cohorts meeting thresholds, with clear fallbacks.
- Before: Carrier follow?ups were reactive. After: Jira tickets auto?opened on score drops with examples and impact.
- Before: Support explained inconsistencies case by case. After: Dashboards showed current quality and affected features.
- Before: Planning re?litigated data trust. After: PMs defended trade?offs with stable, visible quality metrics.
Key Takeaways
- Define data quality explicitly; completeness, timeliness, and consistency should be measured per carrier and lane.
- Validate at ingest; semantic checks and sequence rules prevent noisy events from propagating downstream.
- Gate product features by quality; tie ETAs and alerts to cohorts that meet thresholds and provide safe fallbacks.
- Make quality visible; carrier scorecards and lane heatmaps align Product, Ops, and Support on the same facts.
- Keep humans on exceptions; thresholds, suppression, and variances benefit from review and change control.
- Integrate, dont replace; layer validation and scoring onto your existing warehouse, integrations, and BI.
FAQ
What tools did this integrate with? The pipeline validated and scored events in the existing warehouse (for example, Snowflake), transformed data using the teams preferred tools (for teams using dbt, see dbt docs), and aligned checks to patterns in Great Expectations. Dashboards ran in the companys BI layer, and Jira captured carrier follow?ups and product gates. Event concepts followed a canonical model with optional mapping to GS1 EPCIS.
How did you handle quality control and governance? Validation rules, scoring formulas, and suppression thresholds lived under change control with Product and Data ownership. Changes required review and produced a visible change log. Exceptions and carrier variances captured reason codes and expirations. Lineage connected dashboards back to validations and raw events.
How did you roll this out without disruption? Scoring and suppression ran in shadow mode first, producing dashboards and draft gates while legacy behavior continued. Teams tuned thresholds against known incidents and customer feedback. Gates and suppression were then enabled for selected modes and regions, with fallbacks and manual overrides available early on.
How were the scores calculated and applied? Scores combined dimensions such as milestone completeness, timeliness relative to lane expectations, sequence integrity, identifier consistency, and geo plausibility. Scores rolled up by carrier?lane?service and customer segment. Product gates and dashboards referenced these rollups to enable features and to drive carrier follow?ups.
How did auto?suppression avoid harming customers? Suppression muted unreliable events from driving ETAs and proactive alerts while fallbacks shifted messaging to conservative SLA?based estimates or secondary signals. When suppression engaged, alerts notified Product and Carrier Ops, and Jira tickets captured remediation steps.
How did this handle mixed formats like EDI and APIs? Ingest normalized feeds to the canonical event model, mapping EDI segments and API payloads to common fields. Validation operated on the normalized layer, so checks and scoring applied consistently regardless of source format.
Can this inform carrier negotiations and SLAs? Yes. Scorecards by lane and service provided evidence for SLA discussions, joint improvement plans, and account communications. Because the metrics were visible in product planning as well, they aligned commercial conversations with feature availability and customer promises.
Get a FREE
Proof of Concept
& Consultation
No Cost, No Commitment!


