2020
DOI: 10.1007/978-3-030-66498-5_21
|View full text |Cite
|
Sign up to set email alerts
|

Prototype Selection Using Clustering and Conformance Metrics for Process Discovery

Abstract: Automated process discovery algorithms aim to automatically create process models based on event data that is captured during the execution of business processes. These algorithms usually tend to use all of the event data to discover a process model. Using all (i.e., less common) behavior may lead to discover imprecise and/or complex process models that may conceal important information of processes. In this paper, we introduce a new incremental prototype selection algorithm based on the clustering of process … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 25 publications
0
6
0
Order By: Relevance
“…Clustering event logs can be seen as a preprocessing step to identify sublogs, thus improving the quality of downstream tasks, such as process discovery [17]. Other common encoding methods applied are bags of activities [18], dependency spaces [19], and log footprints [10], for instance.…”
Section: Related Workmentioning
confidence: 99%
“…Clustering event logs can be seen as a preprocessing step to identify sublogs, thus improving the quality of downstream tasks, such as process discovery [17]. Other common encoding methods applied are bags of activities [18], dependency spaces [19], and log footprints [10], for instance.…”
Section: Related Workmentioning
confidence: 99%
“…III. Those models have been discovered from the prototypes of the logs defined by [25] and by using the inductive miner (IM) [22] and the split miner (SM) [21] 3 . We recall that the previous SAT implementation of anti-alignment could not deal with entire large logs due to the complexity of the encoding.…”
Section: Precision Of Real-life Modelsmentioning
confidence: 99%
“…Since its initial adoption, trace clustering has been proposed as an instrument to reduce variability. Discovering process models from clusters, for example, generally improves quality [14]. An early work in the area, presented by Greco et al [15], uses a set of n-grams to encode a trace activity sequence, thus, transforming traces to feature vectors and input clustering techniques.…”
Section: Related Workmentioning
confidence: 99%
“…Trace clustering techniques have been adopted to solve this issue by identifying sub-logs grouped by trace similarity. This way, by detecting groups with homogeneous behavior, process discovery techniques can be executed in these sub-logs, producing higher quality models, which are instead accessible for stakeholders [14]. Trace clustering has also been studied in the context of explainability for PM [19], and, more recently, adapted to incorporate expert knowledge [18].…”
Section: Introductionmentioning
confidence: 99%