Scalable process discovery and conformance checking

Published on May 1, 2018in Science & Engineering Faculty
Sander J. J. Leemans6
Estimated H-index: 6
Dirk Fahland25
Estimated H-index: 25
W.M.P. van der Aalst69
Estimated H-index: 69
Considerable amounts of data, including process events, are collected and stored by organisations nowadays. Discovering a process model from such event data and verification of the quality of discovered models are important steps in process mining. Many discovery techniques have been proposed, but none of them combines scalability with strong quality guarantees. We would like such techniques to handle billions of events or thousands of activities, to produce sound models (without deadlocks and other anomalies), and to guarantee that the underlying process can be rediscovered when sufficient information is available. In this paper, we introduce a framework for process discovery that ensures these properties while passing over the log only once and introduce three algorithms using the framework. To measure the quality of discovered models for such large logs, we introduce a model–model and model–log comparison framework that applies a divide-and-conquer strategy to measure recall, fitness, and precision. We experimentally show that these discovery and measuring techniques sacrifice little compared to other algorithms, while gaining the ability to cope with event logs of 100,000,000 traces and processes of 10,000 activities on a standard computer. © 2016 The Author(s)
  • References (0)
  • Citations (0)
📖 Papers frequently viewed together
37 Citations
25 Citations
2014BPM: Business Process Management
15 Citations
78% of Scinapse members use related papers. After signing in, all features are FREE.
Cited By0