Object-Centric FPR Evaluation
- The paper introduces an object-centric FPR evaluation that quantifies the fraction of false positive context-activity pairs allowed by the model but not observed in logs.
- It explains an algorithmic procedure involving context extraction, event grouping, breadth-first replay, and set comparison for precise FPR calculation.
- The study highlights practical recommendations to mitigate challenges like state-space explosion and log noise through sampling and object abstraction.
Object-centric false positive rate (FPR) evaluation is a rigorous approach to quantifying the quality of process models—specifically object-centric Petri nets—in settings where multiple, interacting case notions are present. Unlike traditional process mining, which relies on a single case notion and sequential events, object-centric frameworks accommodate concurrent, interdependent objects, necessitating more sophisticated metrics for assessing model precision and error rates. In this context, FPR reflects the proportion of possible (context, activity) pairs allowed by the process model that are not actually observed in the event log, thereby indicating the model’s tendency to permit unobserved (false positive) behavior (Adams et al., 2021).
1. Formal Definitions of Object-Centric Precision and FPR
Both the log and the model are characterized in terms of "one-step continuations" under a given context. An object-centric Petri net uses places colored by object types, variable arcs, an initial marking %%%%1%%%%, and a set of final markings. The model’s behavioral language is defined as the set of all pairs where, after replaying some binding sequence , a marking is reached with context , and some enabled transition labeled can fire.
The log language is the set of all pairs such that event in the object-centric event log occurs in the context .
False positives are those (context, activity) pairs allowed by the model but not observed in the log: Precision is then defined as: FPR is the complement, giving: An event-wise average is typically computed, where for each event :
The per-event precision is then , with FPR per-event as its complement, averaged over the event set (Adams et al., 2021).
2. Algorithmic Procedure for FPR Calculation
The recommended computational workflow involves:
- Context Extraction: Compute each event’s context by constructing the event–object graph and extracting multisets of object prefixes as described in Definition 3 of the referenced paper.
- Event Grouping by Context: Aggregate events that share the same context, collating the corresponding visible binding sequences from the log.
- Breadth-First Replay: For each context and its binding sequence, perform a breadth-first replay on the object-centric net. This involves:
- Initializing from , enqueueing .
- While the pending binding list is nonempty, fire all silent transitions exhaustively. For each enabled binding, proceed forward and re-enqueue. When the visible binding list is empty, collect all enabled, visible transitions at each marking.
- Set Comparison: For each context, compare the set of model-enabled activities with the set of log-enabled activities to compute per-context (event) precision and FPR.
Data structures essential for efficiency include:
- Hash maps from contexts to event lists.
- Queues of replay states ().
- Sets to store enabled activities per context.
Computational complexity is for context building in the worst case; replay complexity is exponential in the number of silent loops and combinatorial object-variable bindings (Adams et al., 2021).
3. Illustrative Example
A core exemplification involves the "flight-and-baggage" log () and the discovered net (). For a representative event (“Lift off” for plane p1):
- The computed context comprises prefixes for both plane and baggage objects.
- Replay of this context in the net results in four reachable markings due to silent transitions.
- At each marking, both "Lift off" and "Pick up @ dest" are model-enabled, while only "Lift off" is observed in the log given this context.
- Hence, for , en_L() = {"Lift off"}, en_N() = {"Lift off", "Pick up @ dest"}, yielding per-event precision 0.5, per-event FPR 0.5.
- Aggregation across yields overall fitness 1.0 (complete replayability) and precision ≈ 0.89 (Adams et al., 2021).
4. Interpretability, Sensitivity, and Limitations
This object-centric FPR metric generalizes traditional escaping-edges precision in the single-case setting, admitting the interpretation that the FPR quantifies the fraction of model-allowed continuations not seen in the log. For logs with a single object type and one object per event, the definitions reduce to standard escaping-edges measures.
Notably, contexts that cannot be replayed on the model are skipped in the averaging, which may undermine the reliability of the metric, especially under high log noise or process deviations. A plausible implication is that extensive skipping makes the metric less representative of the model’s overall precision.
Limitations center on state-space explosion from silent transition loops and variable-arc combinatorics, which can render full replay computationally infeasible on large or highly nondeterministic models. The prevalence of generic or re-deployed objects (e.g., equipment used across disparate cases) can create uninformative long-prefix contexts, deflating measured precision.
Approximate replay or random binding sampling is proposed as mitigation, as is object abstraction to avoid overly generic objects. Object-centric alignments for contexts that fail strict replay are suggested as an area for further extension (Adams et al., 2021).
5. Relation to Traditional Quality Metrics
When restricted to a single case notion, the object-centric FPR and precision collapse to their traditional analogues used in process mining, such as the escaping-edges metric. However, object-centric formalism uniquely accounts for multiple, interacting case notions and their induced context dependencies, which are not captured by single-case metrics. This supports fairer and more insightful evaluation of models discovered from real-world processes where multiple entities interact.
6. Practical Recommendations and Best Practices
Effective empirical application of object-centric FPR computation necessitates several practices:
- Pre-filtering or abstraction of ubiquitous objects to prevent artificial context inflation.
- Restriction or sampling of silent transition closures in highly looped models to manage computational demands.
- Supplementing skipped contexts with object-centric conformance alignments where feasible, enhancing metric robustness.
- Transparent reporting of both fitness and precision (or FPR), as well as the proportion of skipped contexts.
These practices allow for sound, reproducible quantification of false-positive behavior in object-centric process mining, enabling principled evaluation and improvement of process discovery techniques (Adams et al., 2021).