Latent Process Discovery Using Evolving Tokenized Transducer

Today organizations capture and store an abundant amount of data from their interaction with clients, internal information systems, technical systems and sensors. Data captured this way comprises many useful insights that can be discovered by various analytical procedures and methods. Discovering re...

Full description

Bibliographic Details
Main Authors: Dalibor Krleza, Boris Vrdoljak, Mario Brcic
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8910574/
id doaj-1e0dcaef28d548d0a46e968ddc55be90
record_format Article
spelling doaj-1e0dcaef28d548d0a46e968ddc55be902021-03-30T00:26:54ZengIEEEIEEE Access2169-35362019-01-01716965716967610.1109/ACCESS.2019.29552458910574Latent Process Discovery Using Evolving Tokenized TransducerDalibor Krleza0https://orcid.org/0000-0001-7350-8858Boris Vrdoljak1Mario Brcic2https://orcid.org/0000-0002-7564-6805Department of Applied Computing, Faculty of Electrical Engineering and Computing, University of Zagreb, Zagreb, CroatiaDepartment of Applied Computing, Faculty of Electrical Engineering and Computing, University of Zagreb, Zagreb, CroatiaDepartment of Applied Computing, Faculty of Electrical Engineering and Computing, University of Zagreb, Zagreb, CroatiaToday organizations capture and store an abundant amount of data from their interaction with clients, internal information systems, technical systems and sensors. Data captured this way comprises many useful insights that can be discovered by various analytical procedures and methods. Discovering regular and irregular data sequences in the captured data can reveal processes performed by the organization, which can be then assessed, measured and optimized, to achieve overall better performance, lower costs, resolve congestions, find potentially fraudulent activities and similar. Besides process discovery, capturing data sequences can give additional behavioral and tendency insights for various observations in the organization, such as sales dynamic, customer behaviour and similar. The issue is that most of the captured data intertwine multiple processes, customers, cases, products in a single data log or data stream. In this article, we propose an evolving tokenized transducer (ETT), capable of learning data sequences from a multi-contextual data log or stream. The proposed ETT is a semi-supervised relational learning method that can be used as a classifier on an unknown data log or stream, revealing previously learned data sequences. The proposed ETT was tested on multiple synthetic and real-life cases and datasets, such as dataset of retail sales sequences, hospital process log involving septic patient treatment and BPI challenge 2019 dataset. Test results are successful, revealing ETT as a prominent process discovery method.https://ieeexplore.ieee.org/document/8910574/Anomaly detectionknowledge acquisitionlearning automatamachine learningpattern recognitionsequences
collection DOAJ
language English
format Article
sources DOAJ
author Dalibor Krleza
Boris Vrdoljak
Mario Brcic
spellingShingle Dalibor Krleza
Boris Vrdoljak
Mario Brcic
Latent Process Discovery Using Evolving Tokenized Transducer
IEEE Access
Anomaly detection
knowledge acquisition
learning automata
machine learning
pattern recognition
sequences
author_facet Dalibor Krleza
Boris Vrdoljak
Mario Brcic
author_sort Dalibor Krleza
title Latent Process Discovery Using Evolving Tokenized Transducer
title_short Latent Process Discovery Using Evolving Tokenized Transducer
title_full Latent Process Discovery Using Evolving Tokenized Transducer
title_fullStr Latent Process Discovery Using Evolving Tokenized Transducer
title_full_unstemmed Latent Process Discovery Using Evolving Tokenized Transducer
title_sort latent process discovery using evolving tokenized transducer
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description Today organizations capture and store an abundant amount of data from their interaction with clients, internal information systems, technical systems and sensors. Data captured this way comprises many useful insights that can be discovered by various analytical procedures and methods. Discovering regular and irregular data sequences in the captured data can reveal processes performed by the organization, which can be then assessed, measured and optimized, to achieve overall better performance, lower costs, resolve congestions, find potentially fraudulent activities and similar. Besides process discovery, capturing data sequences can give additional behavioral and tendency insights for various observations in the organization, such as sales dynamic, customer behaviour and similar. The issue is that most of the captured data intertwine multiple processes, customers, cases, products in a single data log or data stream. In this article, we propose an evolving tokenized transducer (ETT), capable of learning data sequences from a multi-contextual data log or stream. The proposed ETT is a semi-supervised relational learning method that can be used as a classifier on an unknown data log or stream, revealing previously learned data sequences. The proposed ETT was tested on multiple synthetic and real-life cases and datasets, such as dataset of retail sales sequences, hospital process log involving septic patient treatment and BPI challenge 2019 dataset. Test results are successful, revealing ETT as a prominent process discovery method.
topic Anomaly detection
knowledge acquisition
learning automata
machine learning
pattern recognition
sequences
url https://ieeexplore.ieee.org/document/8910574/
work_keys_str_mv AT daliborkrleza latentprocessdiscoveryusingevolvingtokenizedtransducer
AT borisvrdoljak latentprocessdiscoveryusingevolvingtokenizedtransducer
AT mariobrcic latentprocessdiscoveryusingevolvingtokenizedtransducer
_version_ 1724188265656877056