2021
DOI: 10.48550/arxiv.2104.00721
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ProcessTransformer: Predictive Business Process Monitoring with Transformer Network

Zaharah A. Bukhsh,
Aaqib Saeed,
Remco M. Dijkman

Abstract: Predictive business process monitoring focuses on predicting future characteristics of a running process using event logs. The foresight into process execution promises great potentials for efficient operations, better resource management, and effective customer services. Deep learning-based approaches have been widely adopted in process mining to address the limitations of classical algorithms for solving multiple problems, especially the next event and remaining-time prediction tasks. Nevertheless, designing… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0
3

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(19 citation statements)
references
References 29 publications
0
16
0
3
Order By: Relevance
“…The model can be applied to long-distance dependent feature learning and concurrent execution of multiple tasks in the data and can well solve the problems of LSTM in training models, thus improving the model's prediction performance. On this basis, Bukhsh et al [11] proposed the ProcessTransformer model, i.e., modifying the Transformer structure according to the specific process prediction task, to achieve the desired prediction effect.…”
Section: With the Development Of Data Mining And Profound Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…The model can be applied to long-distance dependent feature learning and concurrent execution of multiple tasks in the data and can well solve the problems of LSTM in training models, thus improving the model's prediction performance. On this basis, Bukhsh et al [11] proposed the ProcessTransformer model, i.e., modifying the Transformer structure according to the specific process prediction task, to achieve the desired prediction effect.…”
Section: With the Development Of Data Mining And Profound Learningmentioning
confidence: 99%
“…Based on these mainstream process prediction methods based on RNN and LSTM networks, some studies still use the attention mechanism as an optimization strategy to improve the performance of prediction models. For example, Bukhsh et al [11] proposed the Process Transformer model, that is, to modify the structure of the Transformer network according to specific process prediction tasks to achieve ideal prediction results. Similarly, Wickramanayake et al [28] proposed two types of attention for the prediction task of future activities: the event-level attention to capturing the impact of specific events on the prediction task and the attribute-level attention to reveal which attributes of events affect the prediction task.…”
Section: B Prediction Based On Deep Learningmentioning
confidence: 99%
“…The existing vast majority of PPM techniques can be grouped into three categories according to prediction content, including time prediction [17,40,41], outcome prediction [18,[42][43][44][45], and the next activity (sequences) prediction [19,21,41,[46][47][48]. Moreover, all approaches mentioned above can be grouped into three categories depending on the used techniques, including those based on an extended process model, traditional machine learning, and deep learning.…”
Section: Predictive (Business) Process Monitoringmentioning
confidence: 99%
“…On the basis of GPT-2 that can create a context processed representation by according attention to different parts of the input sentence, the POP-ON model adds a linear network which exploits the event attribute for the residual structure within the Transformer Decoder Layer to reflect elements in the prediction. In this case, the article 7 [80] offers another approach known as the transformer process to solve the problems of changing and complex time-specific data sequences because of multiple control flows in real-life event logs. The main contribution is a process transformer technology for learning high-level representations, which can be utilized to reason through a long sequence of networks like as LSTM from sequence event log data with minimal preprocessing step.…”
Section: Related Workmentioning
confidence: 99%