FAQ

Where should KPI calculation logic live to prevent semantic drift over time?

To prevent KPI semantic drift, KPI calculation logic should live in a single, governed source of truth that all consuming systems reference, rather than being re-implemented in every report, dashboard, or local tool.

Core principle: one governed source of truth

The KPI definition and calculation logic should be owned by a central, controlled layer, with:

  • Clear data model (inputs, filters, exclusions, time windows, aggregation rules)
  • Version control and documented change history
  • Formal change control and validation where required
  • Traceability to requirements, procedures, and standards

Everything else (dashboards, plant views, spreadsheets) should consume these governed KPIs, not recode them.

Practical options for where the logic lives

In regulated, brownfield environments, the central KPI logic commonly sits in one or a combination of:

  • Data warehouse or data lakehouse semantic layer
    KPI logic defined as governed views, metrics, or semantic objects. BI tools query these objects directly instead of writing custom formulas. This works well when you already have an analytics platform and reasonably consistent source data.
  • Dedicated metrics or calculation service
    An application or microservice that takes event/transaction data and returns pre-calculated, versioned KPIs (for example OEE, NPT, COPQ). MES, dashboards, and reports consume these APIs. This can reduce duplication in heterogeneous MES/SCADA landscapes.
  • MES or historian calculation layer
    For shop-floor performance metrics tied tightly to runtime signals, some plants centralize KPI logic in a validated MES/historian layer, then push results to downstream systems. This only works if you can keep that MES layer as the single KPI authority across sites.
  • Governed KPI library or spec repository
    In less integrated environments, KPI logic may be held as SQL scripts, views, or calculation specs in a controlled repository (for example under Git and change control) and reused across tools. This is weaker than a fully central runtime service but still better than ad hoc re-implementation.

What does not work over time is embedding unique KPI logic separately in:

  • Each BI report
  • Each plant-level Excel workbook
  • Each custom integration script
  • Each vendor point solution

That pattern almost guarantees semantic drift as people fix local issues without updating a shared definition.

Key controls to prevent semantic drift

Regardless of the exact technical location, preventing drift depends on governance more than tooling:

  • Authoritative KPI catalog
    Maintain a catalog that defines each KPI, its purpose, inputs, filters, and exact formula. This catalog must match the implemented logic in the metrics layer.
  • Versioned KPI definitions
    Give KPIs explicit versions. When you change a calculation (for example change how planned downtime is treated for OEE), increment the version, document the rationale, and record effective dates.
  • Formal change control
    Route KPI changes through the same change control you use for other critical systems: impact analysis, approvals, test evidence, and deployment records. In regulated settings, treat major KPI logic as configuration under control.
  • Separation of logic from visualization
    BI tools and dashboards should only reference centrally defined metrics or views, not define their own formulas. If a local team thinks they need a variant, it should be added to the central metrics layer, not hand-implemented in a chart.
  • Test suites and regression checks
    Maintain test cases and reference datasets so you can detect unexpected changes in KPI results when data pipelines, MES configurations, or integrations change.
  • Plant- and site-level transparency
    Provide users with a way to see what version of a KPI they are viewing and where it is calculated. This makes it harder for shadow copies to proliferate unnoticed.

Coexistence with existing MES, ERP, and BI tools

In brownfield environments you will usually end up with a hybrid design:

  • Operational systems (MES, historian, SCADA) generate base events and signals (for example machine states, throughput, scrap, alarms).
  • Transactional systems (ERP, QMS, PLM) provide order, material, quality, and cost context.
  • A central metrics or semantic layer combines this data and implements KPI logic under governance.
  • BI tools, plant dashboards, and reports query this layer rather than building KPIs from scratch.

Completely replacing existing MES/ERP or standardizing on a single vendor for KPI logic is rarely feasible in regulated environments, due to qualification and validation burden, downtime risk, and integration complexity. It is usually more practical to:

  • Keep existing systems as data sources.
  • Extract and normalize data into a governed metrics or semantic layer.
  • Gradually refactor local custom KPI logic to call or query that central layer.

During transition, you may have the same KPI calculated both locally and centrally. Use side-by-side comparisons, documented differences, and clear communication of which source is authoritative to avoid confusion.

Minimum viable pattern if you are starting from spreadsheets

If your current reality is heavily spreadsheet-driven, a pragmatic first step is:

  1. Define and document KPI logic in a controlled spec or SQL repository.
  2. Implement that logic as views or calculated fields in a central database or analytics platform.
  3. Point Excel and BI tools to those views instead of maintaining formulas locally.
  4. Introduce basic version control and change approvals for KPI-related views.

This is not as robust as a dedicated metrics service, but it moves the logic out of individual workbooks and into a more governable layer.

Summary

To prevent semantic drift, KPI calculation logic should live in a single, governed metrics or semantic layer that all consuming tools use. The specific technology can vary (data warehouse semantic layer, metrics service, or MES/historian layer), but the non-negotiables are central ownership, versioning, change control, and clear separation between calculation logic and visualization. In mixed-vendor, regulated environments, this usually means adding a governed metrics layer on top of existing systems rather than trying to replace them.

Get Started

Built for Speed, Trusted by Experts

Whether you're managing 1 site or 100, C-981 adapts to your environment and scales with your needs—without the complexity of traditional systems.