To aggregate or to eliminate? Optimal model simplification for improved process performance prediction. (November 2018)
- Record Type:
- Journal Article
- Title:
- To aggregate or to eliminate? Optimal model simplification for improved process performance prediction. (November 2018)
- Main Title:
- To aggregate or to eliminate? Optimal model simplification for improved process performance prediction
- Authors:
- Senderovich, Arik
Shleyfman, Alexander
Weidlich, Matthias
Gal, Avigdor
Mandelbaum, Avishai - Abstract:
- Highlights: A technique for performance-driven model reduction of GSPNs is proposed. The technique relies on foldings that aggregate or eliminate performance information. Foldings preserve model stability and have a bound for the introduced performance estimation error. Given a budget for the estimation error, an optimal sequence of foldings can be found. Abstract: Operational process models such as generalised stochastic Petri nets (GSPNs) are useful when answering performance questions about business processes (e.g. 'how long will it take for a case to finish?'). Recently, methods for process mining have been developed to discover and enrich operational models based on a log of recorded executions of processes, which enables evidence-based process analysis. To avoid a bias due to infrequent execution paths, discovery algorithms strive for a balance between over-fitting and under-fitting regarding the originating log. However, state-of-the-art discovery algorithms address this balance solely for the control-flow dimension, neglecting the impact of their design choices in terms of performance measures. In this work, we thus offer a technique for controlled performance-driven model reduction of GSPNs, using structural simplification rules, namely foldings . We propose a set of foldings that aggregate or eliminate performance information. We further prove the soundness of these foldings in terms of stability preservation and provide bounds on the error that they introduce withHighlights: A technique for performance-driven model reduction of GSPNs is proposed. The technique relies on foldings that aggregate or eliminate performance information. Foldings preserve model stability and have a bound for the introduced performance estimation error. Given a budget for the estimation error, an optimal sequence of foldings can be found. Abstract: Operational process models such as generalised stochastic Petri nets (GSPNs) are useful when answering performance questions about business processes (e.g. 'how long will it take for a case to finish?'). Recently, methods for process mining have been developed to discover and enrich operational models based on a log of recorded executions of processes, which enables evidence-based process analysis. To avoid a bias due to infrequent execution paths, discovery algorithms strive for a balance between over-fitting and under-fitting regarding the originating log. However, state-of-the-art discovery algorithms address this balance solely for the control-flow dimension, neglecting the impact of their design choices in terms of performance measures. In this work, we thus offer a technique for controlled performance-driven model reduction of GSPNs, using structural simplification rules, namely foldings . We propose a set of foldings that aggregate or eliminate performance information. We further prove the soundness of these foldings in terms of stability preservation and provide bounds on the error that they introduce with respect to the original model. Furthermore, we show how to find an optimal sequence of simplification rules, such that their application yields a minimal model under a given error budget for performance estimation. We evaluate the approach with two real-world datasets from the healthcare and telecommunication domains, showing that model simplification indeed enables a controlled reduction of model size, while preserving performance metrics with respect to the original model. Moreover, we show that aggregation dominates elimination when abstracting performance models by preventing under-fitting due to information loss. … (more)
- Is Part Of:
- Information systems. Volume 78(2018)
- Journal:
- Information systems
- Issue:
- Volume 78(2018)
- Issue Display:
- Volume 78, Issue 2018 (2018)
- Year:
- 2018
- Volume:
- 78
- Issue:
- 2018
- Issue Sort Value:
- 2018-0078-2018-0000
- Page Start:
- 96
- Page End:
- 111
- Publication Date:
- 2018-11
- Subjects:
- Generalised stochastic Petri nets -- Model Simplification -- Folding -- Elimination -- Aggregation -- Process Mining
Database management -- Periodicals
Electronic data processing -- Periodicals
Bases de données -- Gestion -- Périodiques
Informatique -- Périodiques
Database management
Electronic data processing
Periodicals
005.7 - Journal URLs:
- http://www.sciencedirect.com/science/journal/03064379 ↗
http://www.elsevier.com/journals ↗ - DOI:
- 10.1016/j.is.2018.04.003 ↗
- Languages:
- English
- ISSNs:
- 0306-4379
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 4496.367300
British Library DSC - BLDSS-3PM
British Library HMNTS - ELD Digital store - Ingest File:
- 11283.xml