Skip to main content
Top
Published in: Journal of Intelligent Information Systems 2/2018

Open Access 24-02-2018

Discovering anomalous frequent patterns from partially ordered event logs

Authors: Laura Genga, Mahdi Alizadeh, Domenico Potena, Claudia Diamantini, Nicola Zannone

Published in: Journal of Intelligent Information Systems | Issue 2/2018

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Conformance checking allows organizations to compare process executions recorded by the IT system against a process model representing the normative behavior. Most of the existing techniques, however, are only able to pinpoint where individual process executions deviate from the normative behavior, without considering neither possible correlations among occurred deviations nor their frequency. Moreover, the actual control-flow of the process is not taken into account in the analysis. Neglecting possible parallelisms among process activities can lead to inaccurate diagnostics; it also poses some challenges in interpreting the results, since deviations occurring in parallel behaviors are often instantiated in different sequential behaviors in different traces. In this work, we present an approach to extract anomalous frequent patterns from historical logging data. The extracted patterns can exhibit parallel behaviors and correlate recurrent deviations that have occurred in possibly different portions of the process, thus providing analysts with a valuable aid for investigating nonconforming behaviors. Our approach has been implemented as a plug-in of the ESub tool and evaluated using both synthetic and real-life logs.

1 Introduction

Modern organizations are usually driven by the processes needed to create and deliver their products and services. Many organizations document their processes using process models. A process model provides a graphical (and often formal) representation of how processes have to be performed. However, in most organizations the behavior prescribed by process models is not enforced by the underlying IT systems. As a result, actual process executions may deviate from the prescribed behavior.
Deviations may be due to several reasons. For example, due to rapid changes in the market or in the organization structure, a process model might become outdated. In this case, deviations indicate that the model should be changed to capture the actual process behavior. Another reason for deviations is to deal with exceptional situations that arise during process executions. For example, an employee may skip some activities to handle an emergency more quickly. These exceptions should be examined to ensure that they conform to guidelines defined for handling these situations. Based on the analysis, the identified exceptions might be integrated into the process model or preconditions for their execution might be defined. Deviations can also indicate that the opportunity for mistakes or even frauds exists. By analyzing these deviations, certain control measures may be defined or the process (and, possibly, the IT system) may be redesigned to prevent the occurrence of undesired behaviors in the future.
To assess the conformance of their business processes, organizations usually employ logging mechanisms to record process executions in event logs and auditing mechanisms to analyze those logs. Event logs consist of traces, each recording data related to the activities performed during a process execution. In last decades, several research efforts have been devoted to the development of automatic and semi-automatic techniques for assisting organizations in the analysis of their event logs to assess the compliance of their processes. Among them, conformance checking (Adriansyah et al. 2011) has been gaining an increasing attention from both practitioners and researchers. Conformance checking techniques usually replay each trace of an event log against a process model to detect possible deviations. State-of-the-art conformance checking techniques are based on the notion of alignment (van der Aalst et al. 2012). Given a trace and a process model, an alignment maps the events of the trace to the ones of a complete run of the model (see (van der Aalst et al. 2012) for a formal definition of alignment). Through this mapping, an analyst can detect possible discrepancies between the observed and the prescribed behavior.
Although a large body of research has addressed issues related to the building of alignments (e.g., Adriansyah et al. 2011; Alizadeh et al. 2015), supporting the analysis of the detected deviations has received little attention so far. Alignments enable a punctual and detailed analysis of single process executions or, at most, of groups of identical executions. However, manual analysis of single executions can likely become challenging, or even unfeasible, when an analyst has to explore a bunch of process executions (e.g., executions of the last six months, the last year...), involving hundreds or thousands of executions. This is especially true when the process under examination involves many possible variants and/or concurrent activities, since single executions will likely differ significantly among each other. We argue that to effectively support process diagnostics further elaboration has to be performed on the detected deviations to provide the analyst with a sort of “deviations dashboard”, reporting analytics and interesting trends regarding the occurred deviations. Indeed, this type of output offers a more high-level perspective on deviant behaviors, from which it is possible to grasp meaningful insights on the deviations that would be otherwise scattered among a multitude of alignments. However, most of the well-known conformance checking techniques typically only provide some basic statistics, like the number of different discrepancies, or the average degree of compliance of process traces to the model.
To address these issues, in a previous work (Genga et al. 2016) we have proposed a framework designed to infer anomalous patterns representing recurrent deviations, together with their correlations. Given an event log and a process model, we applied a frequent subgraph mining technique to extract relevant subgraphs and introduced a conformance checking algorithm to identify the anomalous ones. Then, we proposed to correlate anomalous subgraphs by exploiting frequent itemset algorithms to detect the subgraphs that frequently occurred together and inferring ordering relations among them. Our framework enables an analysis of the deviations that can be considered as orthogonal to the output provided by classic alignment-based techniques. First, instead of determining the discrepancies for each process execution, our approach determines which deviations occurred most frequently. This is motivated by the fact that we could reasonably expect that, in most cases, an analyst would like to ignore very rare deviations and rather focus on those that occur repeatedly and, hence, are more likely envisaged to occur in the future. Moreover, our framework infers portions of process executions involving one or more deviant behaviors, instead of single discrepancies, to provide the most common context(s) of execution within the process in which a given deviations occurred. Finally, the framework allows exploring possible correlations among deviations. It is often the case that a set of low-level deviations are a manifestation of a single high-level deviation (Adriansyah et al. 2013), like the swapping of (sequences of) activities. These higher-level deviant behaviors typically turn out to be much more meaningful for diagnosis than individual low-level deviations. However, in order to diagnose them from a set of alignments, an analyst has to manually investigate the detected (low-level) deviations to reconstruct what happened, which usually turns out to be an error-prone and time-consuming task.
However, the approach in Genga et al. (2016) is devised to derive anomalous patterns from totally ordered traces, with the result that the actual control-flow remains hidden. Extracting anomalous behaviors from sequential traces presents a number of drawbacks. When process executions are represented as totally ordered traces, the same concurrent behavior can be recorded differently in different traces, depending on the activities’ execution time. For instance, given two concurrent activities A and B, we might find totally ordered traces where A was executed before B and other traces where B was executed before A. It is straight to see that recognizing a concurrent behavior from a set of sequential behaviors is far from trivial. As a consequence, given a set of totally ordered traces, one might infer different sequential deviant behaviors that actually correspond to the same deviation. This situation might lead to imprecise diagnoses and even to miss relevant deviant behaviors. Furthermore, it can have a negative impact in terms of both information loss and comprehensibility of the outcome. These issues are as more significant as higher is the degree of concurrency in the model (an example is provided in Section 3.1).
This work extends our previous work to deal with concurrency. We propose an approach for discovering anomalous frequent patterns showing recurrent and correlated deviations modeling both sequential and concurrent behaviors. In particular, we propose to extract frequent anomalous patterns from partially ordered traces, i.e. traces that explicitly represent concurrent behavior among process activities. The main contributions of this work are:
  • We propose a conformance checking algorithm tailored to check the conformance of partially ordered subtraces exhibiting concurrent behavior.
  • We investigate and formalize ordering relations between subgraphs exhibiting concurrent behavior with respect to a process model. Based on the identified relations, we devise an approach to infer partially ordered subgraphs that show correlations among recurrent anomalous behaviors.
  • We propose transformation rules to convert partially ordered subgraphs into anomalous patterns expressed as Petri nets.
  • We present an implementation of the proposed approach as a software plug-in of the Esub tool (Diamantini et al. 2015).
  • We evaluate the proposed approach using one synthetic log and two real-life logs and compare the obtained results with the ones obtained using totally ordered traces (i.e., the approach in Genga et al. 2016).
The remainder of the paper is organized as follows. Section 2 introduces the concepts and notation used through the work. Section 3 provides an overview of the proposed approach along with a motivating example. Section 4 describes our approach for mining relevant subgraphs. Section 5 presents an algorithm for checking the compliance of subgraphs. Section 6 presents an approach to identify partially ordered subgraph and provides guidelines to build anomalous patterns from them. Section 7 presents experiment results. Finally, Section 8 discusses related work and Section 9 draws conclusions and delineates future work.

2 Preliminaries

This section recalls the main concepts used throughout the paper. We start with the concept of process model, which describes the prescribed behaviors of a process. In this work, we use labeled Petri nets to represent process models.
Definition 1 (Labeled Petri Net)
A labeled Petri net is a tuple (P, T, F, A, , mi, mf) where P is a set of places, T is a set of transitions, F ⊆ (P × T) ∪ (T × P) is the flow relation connecting places and transitions, A is a set of labels for transitions representing process activities, : TA is a function that associates a label with every transition in T, mi is the initial marking and mf is the final marking.
In a Petri net, transitions represent activities and places represent states. We distinguish two types of transitions, namely invisible and visible transitions. Visible transitions are labeled with activity names. Invisible transitions are used for routing purposes or represent activities that are not observed by the IT system. Given a Petri net N, the set of activity labels associated with invisible transitions is denoted with InvNA. Multiple transitions can have the same label (i.e., they represent the same activity). Such transitions are called duplicate transitions.
The state of a process model is represented by a marking \(m\in \mathbb {B}(P)\),1 i.e. a multiset of tokens on the places of the net. A process model has an initial marking mi and a final marking mf. A transition is enabled if each of its input places contains at least a token. When an enabled transition is fired (i.e., executed), a token is taken from each of its input places and a token is added to each of its output places. Given a labeled Petri net N, \(m \overset {t}{\rightarrow }_{N} m^{\prime }\) denotes the firing of an enabled transition t from marking m that leads to a new marking m.
Process executions, so-called process instances, are typically recorded in logs. More precisely, the execution of an activity generates an event.
The events related to a process instance are usually ordered in the log. In this work, we adopt partially ordered traces, also referred to as instance graphs in the literature (Lassen and van Dongen 2008). Partially ordered traces explicitly model causal relations and parallelisms among process activities, thus providing a more accurate representation of process behaviors.
Definition 2 (Event, P-Trace, Log)
Let N = (P, T, F, A, , mi, mf) be a labeled Petri net and InvNA the set of invisible transitions in N. An event e represents a unique recorded execution of an activity aA ∖InvN. For an event e, case(e) denotes the process instance in which e occurred and act(e) the activity associated to e. The set of all possible events is denoted by \(\mathcal {E}\). A partially ordered trace (or p-trace) σ is a directed acyclic graph (E, W) where \(E\subseteq \mathcal {E}\) is the set of events related to a single process instance, i.e. for all ei, ejE, case(ei) = case(ej), and W is a set of edges such that (ei, ej) ∈ W indicates that event ei has led to the execution of ej. The edges in W define a partial order2 over E. A log\(\mathcal {L}\) is a set of p-traces.
It is worth noting that although many logging systems record sequential log traces, this does not affect the applicability of our approach. In fact, a number of approaches have been proposed in literature to derive p-traces from sequential log traces (see, e.g., Diamantini et al. 2016; Greco et al. 2005; Hwang et al. 2004; Lu et al. 2014; van Beest et al. 2015). Note that different techniques apply different strategies to infer concurrency among process activities. As a result, the structure of the obtained p-traces and, thus, the structure of the inferred deviant behaviors may vary on the basis of the selected technique. The choice of the employed technique depends on different factors, like the information available on the process, the data stored in the event log, the purpose of the analysis and so on. Nevertheless, the applicability of our approach is not constrained to the technique used to derive the p-traces.
In this work, we use the reachability graph of a Petri net to assess the conformance of p-traces with a process model. We first introduce the notion of marking reachability and then we formally define the notion of reachability graph.
Definition 3 (Marking Reachability)
Let N = (P, T, F, A, , mi, mf) be a labeled Petri net. A marking \(m\in \mathbb {B}(P)\) is one step reachable from a marking \(m^{\prime }\in \mathbb {B}(P)\) if there exists a transition tT such that \(m \overset {t}{\rightarrow }_{N} m^{\prime }\). By extension, marking m is reachable from marking m if there exists a sequence ρ = 〈t1, t2,…, tn〉∈ T such that \(m \overset {t_{1}}{\rightarrow }_{N} m_{1} \overset {t_{2}}{\rightarrow }_{N} {\dots } \overset {t_{n}}{\rightarrow }_{N} m^{\prime }\). We call ρ a firing sequence and, by abuse of notation, we write \(m \overset {\rho }{\rightarrow }_{N} m^{\prime }\) to indicate that firing sequence ρ leads from marking m to marking m.
The set of reachable markings of N, denoted \(\mathcal {R}(N)\), is the set of all markings reachable from the initial marking mi, i.e. \(\mathcal {R}(N)=\{m\mid \exists \rho \in T^{*} : m_{i} \overset {\rho }{\rightarrow }_{N} m\}\).
Note that a marking is reachable by itself with a firing sequence of length 0.
Definition 4 (Labeled Transition System)
A labeled transition system is a tuple (Q,Λ, R, q0) where Q is a set of states, Λ is a set of labels, RQ ×Λ× Q is a set of labeled transitions and q0 is the initial state.
Definition 5 (Reachability Graph)
Let N = (P, T, F, A, , mi, mf) be a labeled Petri net. The reachability graph of N, denoted RG(N), is a labeled transition system (Q,Λ, R, q0) where \(Q= \mathcal {R}(N)\) is the set of markings in N reachable from the initial marking mi, Λ = A, \(R = \{ (m,\ell (t), m^{\prime }) \in \mathcal {R}(N) \times A \times \mathcal {R}(N) | \exists _{t \in T}\ m \overset {t}{\rightarrow }_{N} m^{\prime } \}\) is the set of labeled transitions and q0 = mi.
Intuitively, the reachability graph of a Petri net exhibits all behaviors allowed by the net. In particular, its paths represent the possible firing sequences allowed by the net.

3 Approach

In this section, we discuss the motivations for this work through a running example within the banking domain and provide an overview of our approach.

3.1 Motivating example

Banks typically have internal processes that establish how activities and controls should be performed. Figure 1 shows a process model representing the checking of the profile of a receiver of a money transfer,3 represented as a Petri net. The process starts with the activity Start Receiver Processing (SRP). Then, a pre-profiling phase is executed, starting with activity Start Receiver Pre Profiling (SRPP). This phase involves a set of controls on the receiver’s bank, inter-bank and external profiling (RBPC, RIBPC and REPC respectively). The pre-profiling phase ends with activity Finish Receiver Pre Processing (FRPP). If further checking is required, activity Request Extra Pre Profiling (REPP) is performed. The obtained profile is then evaluated (EPP); on the basis of this evaluation, the process continues either with the management of “high risk” profiles, or with the management of “low risk” profiles. The management of high risk profiles starts with activity Start High Risk Receiver Processing (SHRRP) and involves the execution of three parallel subprocesses. The first subprocess, starting with activity Start High Risk Receiver Profile Check (SHRRPC), involves either an automatic or manual profile check (AHRRPC and MHRRPC) and ends with activity Finish High Risk Receiver Profile Check (FHRRPC). After that, an additional check can be required, represented by activity Request High Risk Receiver Profile Check (RHRRPC). The other two subprocesses involve an analysis of the receiver’s history (HRHA) and a notification to an external authority about the profile that has being processed (HRRAN). At the end of these three subprocesses, the receiver’s request is registered (HRRR) and the high risk profile checking is marked as completed (FHRRP). The processing of low risk profiles starts with activity Start Low Risk Receiver Processing (SLRRP) and involves just an internal and external profile verification (LRIV and LREV, respectively) before registering the receiver (LRRR) and marking the checking phase as completed (FLRRP). The process ends with the execution of activity Finish Receiver Processing (FRP).
Although a process model defines the activities and workflows to be performed, i.e. how processes should be executed, reality may deviate from such a model. Consider, for instance, the case where pre-profiling checking is not performed properly. More precisely, activity FRPP is performed immediately after SRP (i.e., the execution of the activities FRPP and SRPP is swapped), which means that the pre-profiling step has been marked as completed before executing the prescribed controls. This might occur for many reasons. For instance, there might be an implicit work practice of skipping pre-profiling checking when the receiver is an important and trusted customer of the bank; or, it might represent malpractices, leading to a wrong evaluation of the risks associated to the receiver’s profile.
A number of conformance checking techniques have been proposed to assist organizations in the analysis of the observed user behavior against the normative behavior and in the detection of deviations from the specifications. However, these techniques typically focus on the analysis of single process instances. Therefore, these approaches often provide limited insights about deviations and their occurrence, which makes it difficult for an organization to understand and manage them. For instance, if a certain deviation occurs very rarely, it likely corresponds to exceptional circumstances that should be investigated and addressed individually. On the other hand, if it occurs repeatedly, investigating every occurrence of the deviation individually can be time consuming and costly. In these cases, it would be desirable to identify anomalous patterns representing recurrent anomalous behaviors (i.e., deviations from the specification). Anomalous patterns, thus, can provide a baseline for the investigation of deviations requiring systematic and persistent corrective actions that address the source of the problem and, thus, are able to cope with all the occurrences of the identified deviations.
Moreover, to assess the compliance of the observed behavior with the normative behavior, existing conformance checking techniques compare process instances against the process model. However, most of these techniques are only able to deal with sequential traces. A sequential trace is a particular case of p-traces (Definition 2) in which the events in the trace are totally ordered.
Extracting anomalous behaviors from sequential traces, however, presents a number of drawbacks. In particular, the flat representation of sequential traces does not make it possible to represent the control flow of the process and, thus, possible parallelisms between process activities remain hidden. This can have a negative impact both in terms of information loss and in terms of comprehensibility of the outcome by a human analyst. These issues are as more significant as higher is the degree of concurrency in the model.
As an example, consider the swapping of activities FRPP and SRPP discussed above. Figure 2a shows the corresponding p-trace and Fig. 2b shows (a subset of) the sequential traces obtained by flattening the p-trace. If we match each sequential trace against the process model, we obtain different possible instantiations of the anomalous behavior, which are highlighted by a dashed rectangle in the traces of Fig. 2b. Intuitively, these traces differ for the order in which activities RBPC, RIBPC and REPC have been executed. It is up to the analyst to recognize that all of them are related to the same parallel behavior, which can be time-consuming and challenging, especially when the level of concurrency is high. Moreover, sequential instances have an overall support (intended as the percentage of process instances involving the behavior at least once) much lower than the support of the real behavior. As a consequence, when we look for recurrent behaviors, we might miss some (or, in the worst case, all) sequential instances of a parallel behavior.
To address these issues, concurrency has to be taken into account when deriving anomalous behaviors. In our example, considering concurrency (i.e., comparing a p-trace, instead of its corresponding sequential traces, against the model) would make it possible to derive the pattern wrapped by the dashed rectangle in Fig. 2a. It is straightforward to observe that this pattern provides a more accurate and compact description of the anomalous behaviors, thus providing diagnostics that are more meaningful and easy to understand for a human analyst. Moreover, adopting this representation allows us to compute the support of a certain behavior without any information loss, thus providing a more accurate detection of recurrent behaviors.

3.2 Framework

In this work, we propose an approach to mine recurrent anomalous behavioral patterns from historical logging data, explicitly taking into account the process control-flow structure. The approach extends our previous work (Genga et al. 2016) in order to deal with process concurrency. An overview of the approach is given in Fig. 3. It consists of four main phases: (i)relevant subgraph mining, (ii)anomalous subgraph extraction, (iii)subgraph ordering discovery and (iv)pattern design.
Given an event log comprising p-traces, we first apply a frequent subgraph mining technique to extract relevant subgraphs (Section 4). The mined subgraphs can exhibit concurrent executions of activities and are structured in a hierarchy with respect to the inclusion relation occurring among subgraphs. In the anomalous subgraph extraction phase (Section 5), a novel conformance checking algorithm is exploited to extract anomalous subgraphs (i.e., subgraphs that do not fit the prescribed behavior defined in the process model) from the subgraph hierarchy by comparing the mined subgraphs against the process model. In the last two phases (Section 6), we discover ordering relations between anomalous subgraphs with respect to their occurrence in the p-traces and exploit these relations to design anomalous patterns representing frequent anomalous behaviors. In particular, partially ordered anomalous subgraphs are derived by detecting ordering relations among correlated anomalous subgraphs. We finally provide an approach to construct anomalous patterns from partially ordered anomalous subgraphs through transformation rules.
The patterns obtained using the proposed approach can be exploited in various ways. First, they highlight frequent and correlated anomalous behaviors in historical logging data, thus providing a valuable support to diagnostics. They can be used to enhance classic conformance checking techniques to detect high-level deviations, as proposed in Adriansyah et al. (2013). Depending on the context and activities forming the patterns, they can be exploited for supporting the redesign of the process and/or for implementing additional control mechanisms. For instance, a pattern might point out that the process model is outdated. In this case, one may repair the process model to better represent the real processes, as done for example in the work of Fahland and van der Aalst (2015). In other cases, a pattern might represent exceptional behaviors, which should be properly examined. Based on the analysis, these exceptions may also be explicitly modeled in the process model and preconditions can be defined to regulate their execution. When these exceptions represent undesired behaviors, an analysis of the captured patterns can also lead to the definition of control measures or a revision of the process model to prevent the occurrence of errors or frauds in the future. Beside their exploitation at design time, the patterns inferred using our approach can be exploited at run-time by providing a baseline for the definition of monitoring solutions tailored to run-time detection of anomalous behavior in process executions, relieving an analyst from the burden of reevaluating situations already analyzed.

4 Relevant subgraphs mining

The first phase of the approach aims to identify relevant subgraphs occurring in the p-traces recorded in the log. A subgraph γ of a p-trace is a directed acyclic graph formed by a subset of the nodes and a subset of the edges of the p-trace. To determine the set of relevant subgraphs, we apply a Frequent Subgraph Mining (FSM) algorithm to the set of p-traces in the log. The choice of the FSM algorithm depends on the notion of “subgraph relevance” to be used. Existing FSM algorithms often adopt the subgraph support as the underlying notion of relevance. The support of a subgraph is related to its occurrence frequency, which can refer either to all occurrences of the subgraph in a graphs set or to the percentage of graphs involving the subgraph at least once (thus ignoring multiple occurrences of the subgraph in the same graph). In this work, we relate the relevance of a subgraph both to its frequency and size. More specifically, we evaluate the relevance of a subgraph in terms of its Description Length (DL), i.e. the number of bits needed to encode its representation, computed as the sum of the number of bits needed to encode its nodes and the number of bits needed to encode its edges (further details on DL are provided in Holder et al. 1994). Subgraphs with the highest score in terms of DL represent the largest subgraphs that can be obtained at a given frequency. Intuitively, these subgraphs represent core pieces of executions of the process. Extending these subgraphs results in larger, but less frequent, subgraphs representing different possible extensions of the process execution path described by the core subgraphs. To the best of our knowledge, the only FSM algorithm that explicitly considers DL is SUBDUE (Jonyer et al. 2002).
Given a graph set G and a subgraph γ, SUBDUE uses an index based on DL, hereafter denoted by ν(γ, G), which is computed as \(\nu ({\gamma },G)= \frac {DL(G)}{DL({\gamma })+DL(G|{\gamma })}\) where DL(G) is the DL of G, DL(γ) is the DL of γ and DL(G|γ) is the DL of Gcompressed using γ, i.e. the graph set in which each occurrence of γ is replaced with a single node. Intuitively, the relevance of a subgraph is associated with its compression capability: the higher the value of ν is, the lower the DL value of the graph dataset compressed by γ is.
SUBDUE works iteratively. At each step, it extracts the subgraph with the highest compression capability, i.e. the subgraph corresponding to the maximum value of the ν index. This subgraph is then used to compress the graph set. The compressed graphs are presented to SUBDUE again. These steps are repeated until no more compression is possible. The outcome of SUBDUE is a hierarchical structure, where mined subgraphs are ordered according to their relevance, showing the existing inclusion relationships among them. Top-level subgraphs are defined only through elements belonging to input graphs (i.e., nodes and arcs), while lower level subgraphs contains also upper level subgraphs as nodes. Descending the hierarchy, we pass from subgraphs that are very common in the input graph set to subgraphs occurring in a few input graphs.
An example of SUBDUE hierarchy is shown in Fig. 4. Here, we have two top-level subgraphs, γ20 and γ28, and two lower-level subgraphs, γ24 and γ29. The hierarchy shows inclusion relationships between subgraphs. Specifically, γ24 is a child of γ20 because it involves γ20 in its definition. Similarly, γ29 is a child of both γ20 and γ28. In other words, γ24 and γ29 are possible “specializations” of γ20 and of γ20 and γ28 respectively, since they extend the higher level subgraphs with other elements. The hierarchical structure of subgraphs discovered by SUBDUE becomes the input of the next phase of the approach.

5 Anomalous subgraph extraction

The second step of our approach aims to determine among the subgraphs identified in the previous step those that do not fit the given process model. In contrast to most of the existing conformance checking algorithms, the algorithm proposed in this section is tailored to check the conformance of subgraphs corresponding to portions of process executions. The core idea of this algorithm is to replay a subgraph against the given process model represented by its reachability graph.
Given a subgraph γ and the reachability graph RG(N) of a Petri net N, Algorithm 1 determines whether γ is “compliant” or “noncompliant” (i.e., anomalous) with RG(N). First, the algorithm generates all sequences that can be obtained by flattening the subgraph (line 1). Then, each of these sequences is analyzed against RG(N) (lines 2-17). The algorithm uses a stack to store the portion of the reachability graph to be visited. In particular, it stores triples of the form (qi, Qi, seqi) where qi is a visited state in RG(N), Qi is the set of states from which qi can be reached through a sequence of invisible transitions and seqi is the sequence of activities that still need to be compared with RG(N). Note that the set of states Qi is needed to deal with loops formed by invisible transitions, as discussed below.
Given a sequence b, the stack is initialized by inserting a triple (q,{},〈b2,…, bn〉) for each state in RG(N) with an incoming edge labeled with the first activity of b (lines 4-5). Then, at each iteration, Algorithm 1 retrieves a triple from the stack. If no activity remains to be checked, it means that there exists a firing sequence in RG(N) whose labels match with the sequence of activities in b. Thus, the algorithm sets the value of seqFound to true (line 9) and starts comparing the next sequence with RG(N). Otherwise, the algorithm compares the label of outgoing arcs from the current state with the next activity in the sequence (lines 11–15). If they match, the algorithm adds a new triple to the stack for the tail of such an outgoing edge, and the search continues.
Note that the algorithm is robust with respect to the presence of invisible transitions. Edges labeled with invisible transitions are taken into account while exploring the search space, but are not used while matching the paths with the obtained sequences from a subgraph (lines 14–15). To detect loops containing only invisible transitions and avoid exploring them again, the algorithm keeps track of the states from which the current state can be reached through a sequence of invisible transitions. If the next state belongs to the set of states already visited, the algorithm does not explore it again; otherwise, it is explored. If for all the sequences obtained by flattening a subgraph, the algorithm finds a firing sequence in RG(N) whose labels match with the sequence, the subgraph is marked as compliant (line 18); otherwise, it is marked as noncompliant (line 17).
Figure 5 shows an example of application of Algorithm 1. Consider subgraph γ13 in Fig. 5a and (a portion of) the reachability graph corresponding to the Petri net of Fig. 1 in Fig. 5c. The sequences obtained by flattening the subgraph are shown in Fig. 5b. Starting from the first sequence, i.e. seq1, the algorithm looks for the states with an incoming edge whose label matches with the label of the first node in the sequence, i.e. RBPC. There are four states with an incoming edge labeled RBPC, i.e. q2, q5, q6 and q8. Starting from these states, Algorithm 1 checks if there exists a sequence of edges in the reachability graph whose labels matches with the sequence obtained from the subgraph. Consider, for instance, that Algorithm 1 selects q2. The states reached by the algorithm from this state are denoted in gray in Fig. 5c.
We can observe that there is a sequence of edges whose labels match the labels of the second and third nodes in seq1, i.e. 〈(q2, RIBPC, q5),(q5, REPC, q8)〉. Then, q9 can be reached from q8 because (q8, Inv2, q9) is a transition of the reachability graph and Inv2 is a label representing an invisible transition. However, q9 does not have any outgoing edge labeled with SRPP, i.e. the label of the fourth node in seq1. Therefore, the algorithm explores the other initial states (i.e., q5, q6 and q8) to find such a sequence in the reachability graph. As such a sequence does not exist, the subgraph is marked as noncompliant.
It is worth noting that usually inclusion relations exist among the subgraphs returned by subgraph mining techniques. Namely, it is likely to have one or more subgraphs involved in larger subgraphs. These inclusion relations affect the extraction of correlations among deviations (Section 6) and, consequently, the definition of anomalous patterns. Specifically, subgraphs related by an inclusion relation are highly correlated, thus yielding redundant information. A possible way to avoid the presence of inclusion relations among mined subgraphs consists in exploiting special constraints when mining the subgraphs, returning, for instance, only the maximal ones (as done by, e.g. SPIN (Huan et al. 2004) and MARGIN (Thomas et al. 2010)). However, setting constraints based on the subgraphs structure might lead to miss subgraphs describing interesting deviations. A more accurate approach to deal with this issue consists in analyzing the inclusion relations among the mined subgraphs to determine which of them have to be taken into account and which have to be excluded when deriving partially ordered subgraphs to avoid the generation of redundant information.
In this work, we are interested in deriving noncompliant subgraphs that are representative of relevant anomalous behaviors, namely all those subgraphs γi such that i)γi captures a noncompliant behavior, and ii) there does not exists any other subgraph γj such that relevance(γj) > relevance(γi)4 and γjγi. Representative subgraphs represent the smallest and most relevant noncompliant subgraphs. They allow us to define anomalous patterns representing most relevant deviations, without incurring in the redundancy issue previously discussed. Hereafter, we call those subgraphs anomalous subgraphs.
To identify anomalous subgraphs, inclusion relations among noncompliant subgraphs obtained using Algorithm 1 have to be analyzed. To this end, we exploit the hierarchical structure returned by SUBDUE, since it makes inclusion relations explicit. Indeed, by definition every node in the hierarchy is included in all its descendants.
In Genga et al. (2016) we presented an approach to assess the compliance of sequential subgraphs, i.e. subgraphs formed by a sequence of activities, and extract anomalous subgraphs. In this setting, anomalous subgraphs correspond to the minimal noncompliant subgraphs. Indeed, if a subgraph does not comply with the model, its descendants also do not comply with the model, since they also contain the noncompliant sequence of activities represented in the subgraph. Therefore, when a noncompliant subgraph is identified, the algorithm presented in Genga et al. (2016) marks it as anomalous and prunes all the sub-hierarchies rooted in the subgraph.
When considering concurrency, however, this property does not hold. In particular, a subgraph may be marked as noncompliant because it does not exhibit concurrent behavior as defined in the model. However, these subgraphs may be extended by descendants with the required activities. Thus, in presence of concurrency, the descendants of a noncompliant subgraphs can comply with the model. As a consequence, anomalous subgraphs might not correspond to minimal noncompliant subgraphs.
For example, consider the process model in Fig. 1. According to this model, either MHRRPC or AHRRPC should be executed before the execution of FHRRPC. Suppose that, due to the choice construct that involves place p14, events recording the execution of these activities occur less frequently than others and, thus, subgraph γ9 in Fig. 6 is obtained. This subgraph is marked as noncompliant by Algorithm 1 as it it cannot be replayed by the process model. However, the direct descendants of this subgraph, γ10 and γ11, contain an extra node with label MHRRPC and AHRRPC respectively, which makes them compliant with the model.
To deal with these situations, we have extended and revised the approach presented in Genga et al. (2016) to also analyze the descendants of noncompliant subgraphs. If a noncompliant subgraph has some compliant descendants, it means that the subgraph does not capture concurrent behavior as specified in the model. This subgraph is not marked as anomalous but its direct descendants are iteratively analyzed to determine whether they should be marked as anomalous. For example, subgraph γ9 in Fig. 6 is not considered anomalous because all its descendants are compliant. On the other hand, if all the descendants of the subgraph are noncompliant, only the upper-level subgraph is marked as anomalous, since it represents the most representative anomalous behavior compared to its descendants.
This idea is captured by algorithm findAnomalousSubgraphs (Algorithm 2). Given a subgraph hierarchy H and the reachability graph RG(N) of a Petri net N, Algorithm 2 returns the anomalous subgraphs to be considered for the design of anomalous patterns. The algorithm starts by identifying the top-level subgraphs in H (line 1). For each top-level subgraph ts, checkDescendants (Algorithm 3) is used to explore the subgraph hierarchy rooted in ts and identify anomalous subgraphs in the hierarchy (line 4). The union of identified anomalous subgraph is stored in AS and is returned as anomalous subgraphs.
Algorithm 3 takes a subgraph hierarchy H, a subgraph γ and the reachability graph RG(N) of Petri net N as input and returns the set of anomalous subgraphs with respect to the sub-hierarchy of H rooted in γ and RG(N). First, the algorithm checks whether γ is compliant with RG(N) using Algorithm 1 (line 1). Then, the direct descendants of the subgraph are identifies (line 2). If the subgraph has no descendants, depending on whether it is compliant or noncompliant, an empty set (line 5) or the set containing the subgraph is returned (line 7). If the subgraph has some descendants, the algorithm explores the sub-hierarchy rooted in each descendant recursively (line 11) and identifies anomalous subgraphs in each sub-hierarchy. The set of the identified anomalous subgraphs is stored in ASd. If the subgraph is compliant, all descendants marked as anomalous are returned (line 13). Otherwise, the algorithm checks whether all the direct descendants of the subgraph are in ASd (line 15). If this is the case, it means that the subgraph and all of its descendants are noncompliant. Thus, only the subgraph γ, which involves the most representative deviant behavior, is marked as anomalous and, thus, returned to the parent subgraph (line 16). If some descendants of the subgraph are compliant and some are noncompliant, the algorithm returns the noncompliant subgraphs identified by exploring the descendants of the subgraph (line 18).
Figure 7 shows an example of subgraph hierarchy that can be obtained using SUBDUE. In the figure, compliant subgraphs are denoted by a green dash-dotted line rectangle, anomalous subgraphs are denoted by a red full line rectangle, noncompliant subgraphs with some compliant descendants are denoted by a blue dotted line rectangle, and noncompliant subgraphs for which all siblings (with respect to one parent subgraph) are noncompliant, are denoted by an orange dashed line rectangle. The numbers on the top left of rectangles represent the order in which the algorithm explores subgraphs, the set on the top right of rectangles shows the anomalous subgraphs identified after exploring the subgraph and its descendants, the label on the bottom left of rectangles represents whether the subgraph is compliant (C) or noncompliant (NC) with the reachability graph of the Petri net. Subgraphs γ1, γ3 and γ9 are not marked as anomalous subgraphs because some of their descendants are compliant. In this case, their noncompliant descendants, namely γ2, γ5 and γ6, are marked as anomalous subgraphs. As subgraph γ12 and all of its descendants are noncompliant, Algorithm 3 only marks γ12 as an anomalous subgraph in this sub-hierarchy. For the same reason, subgraph γ8 is also marked as an anomalous subgraph. Note that some subgraphs may be reachable from multiple root subgraphs in the subgraph hierarchy. The compliance of these subgraphs is determined with respect to each root subgraph individually. Thus, they are marked as anomalous if they are anomalous with respect to at least one of these root subgraphs. For example, subgraph γ2 is reachable from both subgraphs γ1 and γ8. Algorithm 3 marks γ2 as anomalous because it is anomalous with respect to the hierarchy rooted in subgraph γ1.

6 Subgraph ordering discovery and pattern design

Anomalous behaviors may comprise deviations that manifest in different (and possibly not connected) portions of a process execution. Consider, for instance, the case where an activity a1 that should be executed at the beginning of the process has been swapped with an activity a2 that should be executed at the end of the process. This anomalous behavior will be likely captured as two distinct anomalous subgraphs, one involving a1 and one involving a2. The last two steps of the approach address those situations. In particular, these steps aim to derive ordering relations among anomalous subgraphs describing how apparently different anomalous behaviors are usually correlated, and show how to build anomalous patterns from the discovered partially ordered subgraphs by means of classical process control-flow constructs. In the remainder of this section, we discuss these steps in detail.

6.1 Deriving ordering relations among subgraphs

To infer ordering relations among anomalous subgraphs, we first need to determine which ones tend to occur together. To this end, we generate an occurrence matrix C of the anomalous subgraphs obtained in the previous step of the approach. The occurrence matrix C is an n × m matrix where n is the number of extracted anomalous subgraphs, m is the number of p-traces in the log and each cell cijC represents the number of occurrence of the j-th subgraph in the i-th p-trace.
Determining whether a subgraph occurs in a p-trace corresponds to resolving a subgraph isomorphism problem (Jiang et al. 2013). Formally, given two graphs g = (V, W) and g = (V, W), the goal of subgraph isomorphism is to check if there exists a graph g = (V, W) such that VV, WW and g, g are isomorphic, namely if it is possible to map all the nodes and edges of g on g (involving the corresponding labels). Subgraph g is usually referred as an embedding of g in g. Several algorithms have been proposed in the literature to efficiently compute subgraph isomorphism (see, e.g. Messmer and Bunke 1998; Schmidt and Druffel 1976; Ullmann 1976). In this work, we exploit the subgraph isomorphism technique implemented within the SUBDUE algorithm (Jonyer et al. 2002). Given a subgraph γ and a p-trace σ, such an algorithm returns all the embeddings of γ in σ, thus allowing us to build the occurrence matrix of anomalous subgraphs.
We apply well-known frequent itemset algorithms (Han et al. 2000) to the obtained occurrence matrix in order to derive the sets of subgraphs that co-occur with a support above a given threshold. These sets (hereafter referred to as frequent itemsets) represent the sets of subgraphs that frequently occur together in the p-traces forming the log. Note that frequent itemsets can consist of a single subgraph. These itemsets represent deviant behaviors that occur frequently but for which it is not possible to infer any relation with other deviant behaviors. In particular, they might represent atomic deviant behaviors (i.e., involving a single deviant behavior) or they might actually involve a set of deviant behaviors that tend to occur close to each other during process executions (i.e., few other activities occur in between), with the result that the whole deviant behavior is captured by a single subgraph.
To determine how the subgraphs in a frequent itemset are related to each other, we infer ordering relations between the subgraphs of the itemset in a pairwise fashion. Before defining these relations, we introduce some notation. We first observe that a p-trace can be transformed into a Petri net through graph transformation rules (see Fahland 2008; Ouyang et al. 2009 for examples of transformations from various graph notations into Petri nets). The transformation of a p-trace into the corresponding Petri net is straightforward. Intuitively, the transitions in the Petri net correspond to the nodes (i.e., events) of the p-trace and places correspond to edges. An algorithm for transforming a p-trace into the corresponding Petri net is given in Appendix A. Hereafter, \({\overline {\sigma }}\) denotes the representation of a p-trace σ as a Petri net. Figure 8 shows the Petri net obtained by converting the p-trace of Fig. 2a. For the sake of readability, we sometimes abuse the notation and denote transitions using the activity label of the corresponding events (as done in Fig. 8).
Transforming a p-trace into a Petri net allows us to exploit the reachability graph encoding the behavior of the p-trace to reason on the states (represented as markings) of the subgraphs within the p-trace and, thus, correctly determine ordering relations between the subgraphs occurring in the p-trace. Let σ = (E, W) be a p-trace and γ = (E, W) a subgraph of σ (with EE and WW). We denote \(M_{ini}^{{\sigma }}({\gamma })\) the set of markings in \(\mathcal {R}({\overline {\sigma }})\) that enable the firing of an event in γ. Intuitively, \(M_{ini}^{{\sigma }}({\gamma })\) is the set of markings reachable through a firing sequence from the initial marking mi of \(\overline {\sigma }\) such that it does not includes any event of γ and the next event that can be fired belongs to γ. Formally, \(M_{ini}^{{\sigma }}({\gamma })=\{m\mid \exists \rho \in E^{*} : m_{i}\overset {\rho }{\rightarrow }_{{\overline {\sigma }}}m \wedge \nexists e\in E^{\prime } : e\in \rho \wedge \exists m^{\prime }\in \mathcal {R}({\overline {\sigma }}) : m\overset {e_{i}}{\rightarrow }_{{\overline {\sigma }}}m^{\prime } \wedge e_{i}\in E^{\prime }\}\). Similarly, \(M_{end}^{{\sigma }}({\gamma })\) denotes the set of markings in \(\mathcal {R}({\overline {\sigma }})\) reached by firing all events in γ. Formally, \(M_{end}^{{\sigma }}({\gamma })=\{m\mid \exists \rho =\langle e_{1},\dots ,e_{n}\rangle \in E^{*} : m_{i}\overset {\rho }{\rightarrow }_{{\overline {\sigma }}}m \wedge \forall e\in E^{\prime }\ e\in \rho \wedge e_{n}\in E^{\prime }\}\), i.e. \(M_{end}^{{\sigma }}({\gamma })\) is the set of markings reachable through a firing sequence ρ from the initial marking mi of \(\overline {\sigma }\) that includes all the events of γ with the additional constraint that the last event of ρ belongs to γ.
We now have the machinery necessary to define ordering relations between subgraphs with respect to a p-trace. Let (γi, γj) be an ordered pair of subgraphs occurring in a p-trace σ.5 We can derive one of the following ordering relations for (γi, γj):
Strictly sequential
γj is strictly sequential to γi, denoted as γisseqγj, iff (i)\(M_{end}^{{\sigma }}({\gamma }_{i}) = M_{ini}^{{\sigma }}({\gamma }_{j})\) and (ii)\(|M_{end}^{{\sigma }}({\gamma }_{i})| = | M_{ini}^{{\sigma }}({\gamma }_{j})| = 1\), where |X| denotes the cardinality of a set X. Intuitively, the strictly sequential relation states that events corresponding to the nodes of subgraph γj will occur immediately after all the events corresponding to the nodes of subgraph γi have occurred. Specifically, (i) imposes that events in γj are enabled by the firing of all events in γi, thus guaranteeing that the events in γj can occur only after all the events in γi have occurred; (ii) guarantees that the next event that can be observed after the firing of all events in γi belongs to γj.
Sequential
γj is sequential to γi, denoted as γiseqγj, iff (i) for every marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) there exists a marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\) such that m is reachable from m,6(ii) for every marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\) there exists a marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) such that m is reachable from m, (iii)\(M_{end}^{{\sigma }}({\gamma }_{i}) \cap M_{ini}^{{\sigma }}({\gamma }_{j}) \neq \emptyset \) and (iv)\(|M_{ini}^{{\sigma }}({\gamma }_{j})|>1\) or \(|M_{end}^{{\sigma }}({\gamma }_{i})|>1\). Intuitively, (i) and (ii) guarantee that events in γj can only occur after all events in γi have occurred; (iii) guarantees that there exists some state reached by firing all events in γi in which events in γj are enabled. It is worth noting that (i) of the strictly sequential relation implies these three conditions. However, differently from the strictly sequential relation, the sequential relation allows some events (not belonging to any of the two subgraphs) to occur between the events in γi and the events in γj. This is captured by (iv).
Eventually
γj eventually occurs after γi, denoted as γievγj, iff (i)\(M_{end}^{{\sigma }}({\gamma }_{i})\cap M_{ini}^{{\sigma }}({\gamma }_{j}) = \emptyset \) and (ii) there exists a marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) that is reachable from a marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\). Intuitively, the eventually relation states that an event in γj can only occur after all events in γi have occurred and an arbitrary number of events (at least one) must occur between the two subgraphs. In particular, we can observe that (i) and (ii) guarantee that an event in γj can only occur after all events in γi have occurred, i.e. for every marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) there exists a marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\) such that m is reachable from m and for every marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\) there exists a marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) such that m is reachable from m. (Note that these conditions corresponds to (i) and (ii) of the sequential relation.) Moreover, the disjointness of \(M_{end}^{{\sigma }}({\gamma }_{i})\) and \(M_{ini}^{{\sigma }}({\gamma }_{j})\) guarantees that every firing sequence from a marking \(m \in M_{end}^{{\sigma }}({\gamma }_{i})\) to a marking \(m^{\prime }\in M_{ini}^{{\sigma }}({\gamma }_{j})\) has at least length 1, i.e. at least one event has to occur after all events in γi have occurred in order to enable any event in γj.
Interleaving
γi and γj are interleaving, denoted as γiintγj, iff there exist a marking \(m\in M_{ini}^{{\sigma }}({\gamma }_{j}) \setminus M_{end}^{{\sigma }}({\gamma }_{i})\) and a marking \(m^{\prime }\in M_{end}^{{\sigma }}({\gamma }_{i})\) such that m is reachable from m. Intuitively, the interleaving relation captures all cases where an event in γj can occur before all the events corresponding to the nodes of the subgraph γi have occurred. For instance, interleaving relations capture cases where the two subgraphs contain events that can occur concurrently and cases where the subgraphs share some nodes (i.e., the two subgraphs are overlapping (Genga et al. 2016)).
To derive the relations between the subgraphs in a frequent itemset, we assess the ordering relations between their embeddings pairwise. Specifically, for each ordered pair of embeddings (γi, γj) of the subgraphs occurring in an itemset, we analyze the reachability from the set of markings resulting from the execution of γi (i.e., Mend(γi)) to the set of markings enabling the execution of γj (i.e., Mini(γj)) in the reachability graph corresponding to (the Petri net representation of) the p-traces in which the itemset occurs. Note that a different type of relation can hold for (γi, γj) and (γj, γi). In particular, it is easy to observe that, if the relation for pair (γi, γj) is strictly sequential, sequential or eventually, the relation for (γj, γi) is interleaving. In such cases, we discard pair (γj, γi) and, thus, the interleaving relation from the analysis as it is not representative of the ordering relations between the two subgraphs. Moreover, it is worth noting that, in the presence of noisy logs (and of multiple embeddings of the same subgraph), we can detect unreliable relations. To deal with this issue, we evaluate the occurrence frequency of ordering relations within the p-traces in which the itemset occurs and only consider those ordering relations whose occurrence frequency is above a given threshold.
As an example, consider the p-trace represented as a Petri net in Fig. 8 and frequent itemset {γ12, γ18, γ21}. The embeddings of each subgraph are delimited by a dotted (γ12), a dot-dashed (γ18) and a dashed (γ21) line in the figure (as there is only one embedding for each subgraph in the itemset, hereafter we do not distinguish a subgraph from its embeddings.) To define the ordering relations among these subgraphs in the p-trace, we need to analyze the reachability graph of the p-trace. Figure 9 shows a portion of the reachability graph corresponding to the Petri net in Fig. 8. For the sake of space, we only consider the relevant cases. Let us start with the ordered pair of subgraphs (γ12, γ18). To determine the ordering relation between these two subgraphs, we have to compare the set of marking reached after γ12 is executed (i.e., \(M_{end}^{{\sigma }}({\gamma }_{12})\)) with the set of markings enabling the execution of activities in γ18 (i.e., \(M_{ini}^{{\sigma }}({\gamma }_{18})\)). From Fig. 9 we can observe that \(M_{end}^{{\sigma }}({\gamma }_{12}) =\{[p_{8}]\}\) and \(M_{ini}^{{\sigma }}({\gamma }_{18})=\{[p_{8}]\}\). Hence, we have γ12sseqγ18. Intuitively, this relation states that the events in γ18 are enabled when all events in γ12 have occurred. This is also immediate by observing the Petri net in Fig. 8. Note that, since we have identified a strictly sequential relation for the ordered pair (γ12, γ18), we do not need to check the ordered pair (γ18, γ12).
Consider now the ordered pair of subgraphs (γ12, γ21). To determine the ordering relation between these two subgraphs, we have to compare the set of marking reached after γ12 is executed (i.e., \(M_{end}^{{\sigma }}({\gamma }_{12})\)) with the set of markings enabling the execution of the activities in γ21 (i.e., \(M_{ini}^{{\sigma }}({\gamma }_{21})\)). From Fig. 9, we can observe that \(M_{ini}^{{\sigma }}({\gamma }_{21})=\{[p_{10},p_{11}],[p_{11}, p_{12}]\}\). Since there exists a firing sequence from marking [p8] to every marking in \(M_{ini}^{{\sigma }}({\gamma }_{21})\) and the length of such paths is greater than 1, we derive γ12evγ21. Note that \(M_{ini}^{{\sigma }}({\gamma }_{21})\) includes marking [p10, p11] besides marking [p11, p12]. Although this may appear counterintuitive by looking at the Petri net in Fig. 8, in presence of concurrency, it should be taken into account that activities outside a subgraph can be executed in parallel to the activities of the subgraph. Specifically, we can observe in Fig. 9 that marking [p10, p11] enables the firing of an activity in γ21, i.e. \([p_{10}, p_{11}]^{\underrightarrow {\mathit {\,\,\,LRERV\,\,\,}}} [p_{10}, p_{13}]\) is allowed by the model. We stress that ignoring marking [p10, p11] results in discarding behaviors involving subgraph γ21, which can lead to identify the wrong relation between subgraphs.
Finally, let us consider the ordered pair (γ18, γ21). To determine the relation between these two subgraphs, we have to consider the relation between the markings in \(M_{end}^{{\sigma }}({\gamma }_{18})\) and \(M_{ini}^{{\sigma }}({\gamma }_{21})\). As discussed above, \(M_{ini}^{{\sigma }}({\gamma }_{21})=\{[p_{10},p_{11}],[p_{11}, p_{12}]\}\); moreover, we can observe in Fig. 9 that \(M_{end}^{{\sigma }}({\gamma }_{18})=\{[p_{11},p_{12}],[p_{12},p_{13}]\}\). From Fig. 9, it is easy to observe that there exists a firing sequence from a marking in \(M_{ini}^{{\sigma }}({\gamma }_{21})\setminus M_{end}^{{\sigma }}({\gamma }_{18})\) to a marking in \(M_{end}^{{\sigma }}({\gamma }_{18})\). Specifically, firing sequence ρ = 〈LRERV, LRIRV〉 leads from marking [p10, p11] to marking [p12, p13], i.e. \([p_{10},p_{11}]\overset {\rho }{\rightarrow }[p_{12},p_{13}]\). Therefore, γ18intγ21. Once again, we remark that ordering relations between subgraphs in an itemset should be inferred based on their markings within the p-trace (Fig. 9) rather than on the basis of the configuration of places in the Petri net representing the p-trace (Fig. 8). In fact, by looking at Fig. 8, one may infer a sequential relation for the ordered pair (γ18, γ21). This, however, captures only a particular behavior of the subgraphs within the p-trace. On the other hand, it is evident from the reachability graph in Fig. 9 (which explicitly represents the behavior of the p-trace) that, in general, some activities in γ18 and γ21 can be executed concurrently and, therefore, the two subgraphs are interleaving.
The identified ordering relations between subgraphs of a frequent itemset can be combined to define the partially ordered subgraph for the itemset. Figure 10 represents the partially ordered anomalous subgraph obtained by combining the ordering relations found for frequent itemset {γ12, γ18, γ21}. Note that the support of a partially ordered subgraph is always smaller than or equal to the support of the frequent itemset from which it is derived. Moreover, it depends on the support of the ordering relations among these subgraphs. In particular, the support of a set of partially ordered subgraphs is at most equal to the percentage of traces in which the less frequent of its ordering relations holds.

6.2 Pattern design

Partially ordered subgraphs represent how anomalous subgraphs that frequently occur together are correlated. However, in order to exploit them for improving the process model and/or for conformance checking, partially ordered subgraphs have to be transformed into a more suitable representation. In this section, we present a number of transformation rules to construct anomalous patterns from partially ordered subgraphs.
To provide a formal representation of anomalous patterns, we model them as Petri nets. Moreover, we use the notion of Ω-transition introduced in Ramezani et al. (2012) to ensure the applicability and reuse of anomalous patterns. Intuitively, Ω-transitions are used to mimic the occurrence of any other transition than the ones expected by the pattern in a given state. This makes it possible to abstract the patterns from events whose occurrence is not representative of the corresponding anomalous behavior.
Figure 11 represents the anomalous patterns obtained by two anomalous subgraphs, γi and γj, connected through a strictly sequential, a sequential and an eventually relation where \(\overline {{\gamma }_{i}}\) and \(\overline {{\gamma }_{j}}\) denote the Petri net representation of γi and γj respectively.7 The pattern encoding the strictly sequential relation (Fig. 11a) shows that the activities in γj should occur immediately after the activities in γi have been executed in order for the pattern to be successfully executed. Similarly, the pattern encoding the sequential relation (Fig. 11b) is successfully executed if the activities in γj occur after the activities in γi; however, in this case, other activities may be executed between the activities in γi and the activities in γj as represented by the Ω-transition connected to place p2. On the other hand, in order for the pattern encoding the eventually relation (Fig. 11c) to be successfully executed it is required that at least one activity is executed between the activities in γi and the activities in γj. This is modeled in Fig. 11c by the Ω-transition between places p2 and p3 and the Ω-transition connected to place p3. Note that the initial place p1 of these patterns is connected to a Ω-transition. This structure is used to model the case where a pattern can occur in an arbitrary position within the p-trace.
The construction of anomalous patterns involving interleaving relations requires more attention. The interleaving relation can represent different situations, e.g. it can indicate that two subgraphs are (partially) concurrent or that they overlap on one or more activities. We abstract from the specific situation at hand by merging the two subgraphs involved in the relation into a new subgraph that encompasses the behaviors represented by both subgraphs and their interconnections. Let γi = (Ei, Wi) and γj = (Ej, Wj) be two subgraphs in an interleaving relation and S = {σ1σn} a set of partially ordered traces, each involving at least one embedding for each subgraph. The subgraph obtained by merging γi and γj is \({{\gamma }_{k}}=(E_{i} \cup E_{j}, W_{i} \cup W_{j} \cup \{(e,e^{\prime }) \mid ((e \in W_{1} \land e^{\prime } \in W_{2}) \lor (e^{\prime } \in W_{1} \land e \in W_{2})) \land \forall _{\sigma _{m}=(E_{m}, W_{m}) \in S}((e,e^{\prime }) \in E_{m}) \} )\). Intuitively, the set of nodes of γk is obtained by the union of the sets of nodes of γi and γj. Similarly, the set of edges of γk is obtained by the union of the sets of edges of γi and γj. In addition, the set of edges of γk includes the edges that connect nodes of γi to nodes of γj and vice versa (if any). The existence of these edges can be determined by observing how the embeddings of γi and γj are related in the p-traces in S.
An interleaving relation, however, does not indicate whether other activities/subgraphs can be executed between the subgraphs involved in the relation or not. To distinguish these cases, we propose two anomalous patterns for interleaving relations. The first pattern (Fig. 12a) aims to model situations in which no other activities occur in parallel to γk; the second pattern (Fig. 12b) captures the cases where other activities/subgraphs can be executed concurrently. In the figures, \(\overline {{\gamma }_{k}}\) denotes the Petri net representation of γk. As for the patterns in Fig. 11, both patterns are preceded by a Ω-transition, representing that an arbitrary number of other activities can be executed before the activities in γk. However, the pattern in Fig. 12a represents that no other activities/subgraphs can occur concurrently; once an activity in γk has been fired, nothing else can be executed until the pattern is completely executed. On the other hand, the pattern in Fig. 12b allows for the concurrent execution of an arbitrary number of other activities before, during and after the execution of γk. This is modeled by the Ω-transition in parallel to γk. One might argue that the pattern in Fig. 12b is included in the pattern of Fig. 12a. Although this is indeed the case, the pattern in Fig. 12a is more precise and should be preferred when it is known that nothing can occur in parallel with γk, for instance, to prevent false positives when the pattern is used for conformance checking. Note that it is possible to automatically select the most suitable choice for the pattern at hand by checking the presence of activities executed in parallel to the activities of the partially ordered subgraphs in the traces in which these subgraphs occur.
The transformation rules presented in Figs. 11 and 12 show how two subgraphs connected by an ordering relation can be combined. However, partially ordered subgraphs can encompass more than two anomalous subgraphs. In this case, the ordering relations between the subgraphs forming a partially ordered subgraph can “interfere” with each other and, thus, user intervention may be required to design the corresponding anomalous pattern. In particular, user intervention may be required to determine in which order the subgraphs should be combined in order to construct the pattern that better fits the process. This is especially true in the presence of interleaving relations, where subgraphs are merged. As an example, consider the partially ordered subgraph in Fig. 10. We can observe in the figure that subgraphs γ12 and γ21 are connected by the eventually relation, suggesting that the pattern in Fig. 11c could be used. However, subgraph γ12 is connected to subgraph γ18 through a strictly sequential relation, and subgraphs γ18 and γ21 are interleaving. This may indicate that the Ω-transition between places p2 and p3 in Fig. 11c can be replaced by some events in γ18, depending on the process model. Figure 13 shows the pattern that can be obtained from the partially ordered subgraph in Fig. 10.8
The proposed approach supports an analyst in the exploration of the different patterns obtained by selecting and merging different pairs of subgraphs. In particular, by automatically generating possible alternatives to combine pairs of subgraphs, it is possible to provide a valuable aid to the analyst in selecting the best way of combining the entire set of subgraphs. The definition of a systematic approach for the (semi)automated generation of anomalous patterns from partially ordered subgraphs involving more than two subgraphs, however, requires more investigation. The design of such an approach is left for future work.

7 Experiments

We have implemented our approach as two modules of the Esub tool (Diamantini et al. 2015), which extend the Anomalous Subgraphs Checking and Partial Order Discovery modules presented in Genga et al. (2016) to deal with concurrency. The tool can be found at http://​kdmg.​dii.​univpm.​it/​?​q=​content/​esub and a demonstration of the tool in Genga et al. (2017). The first module implements steps 1 and 2 of the approach. It takes as input (i) a set of p-traces and (ii) the reachability graph of a Petri net, and uses SUBDUE to generate the subgraph hierarchy. Then, the module applies Algorithm 2 to extract anomalous subgraphs. Figure 14a shows a screenshot of the module displaying a portion of the hierarchical structure derived by SUBDUE, indicating which subgraphs are anomalous (see Section 5 for an explanation of the meaning of the colors). The second module implements the third step of the approach; it takes as input the set of frequent itemsets and derives the partially ordered anomalous subgraphs. A screenshot of this module is shown in Fig. 14b. Each edge is labeled with the type of relation it represents. Solid lines are used for sequential and strictly sequential relations, dashed lines for eventually relations and dotted double-arrowed lines for interleaving relations.
To evaluate the approach, we performed a number of experiments on synthetic datasets and two real-world event logs. The aim of the experiments on synthetic data is to perform controlled experiments and assess the accuracy of the approach. We used real-life case studies to show that the approach provides useful insights and is robust to logs and models with real-life complexity. Both studies are reported in the following subsections.

7.1 Synthetic datasets

Settings
For the experiments with synthetic data, we designed the Petri net modeling the bank transaction process in Fig. 1 using CPN tools (http://​cpntools.​org/​) and generated two event logs, used in two set of experiments. Each log consists of 2000 p-traces but involves a different number of events, i.e. 39014 events and 39035 events. This difference is due to the different kind of deviations we inserted in each event log to assess the capability of the approach in detecting recurrent anomalous behaviors, as explained below.
In a first set of experiments, we have manipulated the generated p-traces by inserting a number of deviations, namely swaps, repetitions and replacements. A swap occurs when two or more activities are executed in the opposite order compared to the order defined by the model; we swapped the execution of activity 〈SRPP〉 with the one of 〈FRPP〉, and the execution of 〈LRRR〉 with the execution of 〈FLRRP〉. A repetition means that a given (sequence of) activity(ies) is repeated multiple times (without belonging to a loop); we added two repetitions, namely the repetition of REPP and of LRERV. Finally, a replacement indicates that a given (sequence of) activity(ies) is executed instead of another one; in our experiments, activity SHRRP was replaced with sequence 〈SRP, FRP〉, and 〈HRRR, FHRRP, FRP〉 with 〈EPP, SLRRP〉. For each deviation, we set a probability of occurrence of 30%. By doing so, we obtain an event log involving several varied anomalous behaviors, where it is however possible to recognize some regularities.
To test the robustness of the approach in the presence of noise, in a second experiment we randomly added/removed process activities in portions of the process not involved by the occurrence of the deviations introduced for the previous experiments. The amount of noise added to a trace is equal to the 10% of the length of the trace. By doing so, we obtain an event log significantly affected by random noise. This can be considered an extreme situation as this amount of noise is much higher than what is typically expected in a real-world context.
For the experiments, we filtered out fitting traces as we are interested in the detection of recurrent anomalous behavior. Table 1 summarizes the event logs used for the experiments. The table reports the number of non-fitting traces, the total number of events and the average/minimum/maximum number of events per trace.
Table 1
Event logs used for the synthetic experiments
Experiment
Non-fitting traces
Events
Avg. events per trace
Min events per trace
Max events per trace
No random noise
1500
30238
20
13
41
With random noise
1995
38963
20
12
41
To extract the relevant subgraphs from the logs we used the traditional SUBDUE implementation available at http://​ailab.​wsu.​edu/​subdue/​, in its default configuration.9 To derive partially ordered subgraphs, we first used FP-Growth (Han et al. 2000) to mine all co-occurring subgraphs with a minimum support threshold of 5%. We point out that this can be considered a good support value in our experiments. In fact, the frequency assigned to the deviations, together with the complex structure of the process model involving choice constructs, implies that we cannot expect to detect patterns with a very high support value. This reflects what we reasonable expect to find in a real-world context, where it is unlikely to detect very frequent anomalies (unless the process model is outdated). Ordering relations were inferred using a threshold of 50%.
It is worth noting that partially ordered subgraphs may have a support smaller than the 5% threshold set for the frequent itemsets. The support of a partially ordered subgraph depends both on the support of its itemset (that represents its upper bound) and on the support of the ordering relations connecting the subgraphs in the itemset. For instance, consider an event log consisting of 100 traces and an itemset involving two anomalous subgraphs γi, γj, whose support is 10%. Let us assume that an analysis of traces in which the itemset occurs shows a sequentially relation between the two subs in 90% of the cases and a strictly sequentially relation in the remaining ones. By setting a threshold for ordering relations equals to 50%, only the sequentially relation is considered in the partially ordered subgraph. It is easy to observe that the support of the resulting partially ordered subgraph is 9%.
To show the advantages of considering concurrency in the discovery of recurrent anomalous behavior, we compared the results obtained by the proposed approach with the results obtained using the approach in Genga et al. (2016), that relies on sequential traces for the discovery of recurrent anomalous behavior. For the comparison, we considered the following criteria:
  • Number of mined subgraphs (Subgraphs);
  • Number of anomalous subgraphs (Anomalous Subgraphs);
  • Number and average support of partially ordered subgraphs (P.O., Average support P.O. respectively);
  • Average number of anomalous subgraphs per partially ordered subgraphs (Average subgraphs per P.O.);
  • Average number of activities per partially ordered subgraphs (Average activities per P.O.).
These criteria provide an indication of the quality of diagnostic information that can be obtained using both the approaches. For instance, the average number of anomalous subgraphs forming a partially ordered subgraphs indicates the ability of the approach to correlate anomalous behaviors occurring in (possibly) different portions of a process instance. On the other hand, the average number of activities in a partially ordered subgraph provides an indication of the complexity of the anomalous behaviors that can be captured by the approach.
Results
Table 2 shows the results of our experiments with synthetic data. For the first set of experiments (no random noise), we obtained 329 anomalous subgraphs and 25 partially ordered subgraphs (P.O.) using p-traces. Partially ordered subgraphs have an average support of 13.22%. The analysis of the obtained partially ordered subgraphs shows that all the inserted deviations have been detected by the approach. It is worth noting that although the total number of subgraphs mined from p-traces was significantly lower than the number of subgraphs mined from sequential traces (628 versus 1945), we found much more anomalous subgraphs in the first case (i.e., 329 versus 22). This can be explained by recalling that when considering sequential traces, children of anomalous subgraphs are always anomalous; thus, they are pruned during the conformance checking phase. On the other hand, when concurrency is involved, children of anomalous subgraphs should be analyzed as well, as explained in Section 3.
Table 2
Results for synthetic datasets
Experiment
Trace type
Subgraphs
Anomalous subgraphs
P.O.
Average support P.O.
Average subgraphs per P.O.
Average activities per P. O.
No random noise
p-trace
628
329
25
13.22%
1.48
8.24
 
sequential
1945
22
13
11.87%
1
4.76
With random noise
p-trace
1117
485
657
4.85%
3.49
14.02
 
sequential
5130
618
30
10.93%
1.46
3.43
It is worth noting that we found significantly more complex (and, hence, potentially more interesting) patterns when using p-traces, as shown by the average number of subgraphs and activities per partially ordered subgraphs. In particular, in the noise-free setting we found partially ordered subgraphs involving on average 1.46 and 8.24 activities when using p-traces; while all partially ordered subgraphs mined from sequential traces consist of only one subgraph and involve, on average, 4.76 activities. This is due to the presence of parallel behaviors in the process. In the experiment with the sequential traces, we have multiple sequential instantiations of parallel behaviors, each of them with a support corresponding to a fraction of the support of the behavior they belong to. It is straightforward to see that this affects the detection of correlations between different deviations.
The results obtained from the experiments with random noise confirm the insights we gained from the noise-free ones. We can observe that the insertion of random noise led to the generation of a much larger number of subgraphs. Since noise can be seen as a form of nonconformance, a significantly larger amount of anomalous subgraphs and of partially ordered subgraphs were obtained in these experiments, as shown in Table 2. In particular, the presence of noise led to generate subgraphs that are smaller on average than the ones obtained in the previous experiments. This can be easily observed in Table 2 by computing the ratio of the average number of activities in partially ordered subgraphs over the average number of subgraphs in partially ordered subgraphs. Obtaining smaller subgraphs may indicate that, while in the previous experiments we could mine subgraphs representing entire deviations, in the experiments with random noise the inserted deviations can be scattered among different subgraphs. Therefore, the ability of correlating anomalous behaviors is crucial to reconstruct deviations.
An analysis of the results of the experiments with random noise shows that all the inserted deviations can be detected using both p-traces and sequential traces. However, we obtained more complex and meaningful patterns when using p-traces, both as regards the average number of subgraphs (3.49 versus 1.46) and as regards the average number of activities (14.02 versus 3.43). This difference, as well as the gap between the number of partially ordered subgraphs mined from p-traces and from sequential traces (657 versus 30), is mainly due to the presence of concurrent behaviors, as for the previous experiments. Indeed, 17 of the partially ordered subgraphs mined from sequential traces involve just one subgraph, while the remaining involve at most two subgraphs, with the exception of a single partially ordered subgraph that involves three subgraphs. This provides a clear evidence of the difficulties encountered in deriving correlations among deviations when process control-flow is not taken into account. It is worth noting that the higher complexity of patterns mined from p-trace also explains the gap arisen between the average support of partially ordered subgraphs mined from p-traces and sequential traces (4.85% versus 10.93%). Indeed, larger a partially ordered subgraph is and less likely is that a high number of its embeddings occur, especially in the noisy setting of our experiments. Indeed, it is reasonable to expect large subgraphs to have lower support than small subgraphs. Recall that in our experiments we introduced random noise in the log by randomly adding/removing process activities in some portions of the process. Therefore, larger a subgraph is, more likely is that the subgraph involves those activities. This decreases the possibilities of finding a high number of embeddings of these subgraphs and, thus, their support is generally low.
A qualitative analysis of the partially ordered subgraphs obtained in the experiments shows that all inserted deviations have been captured using both p-traces and sequential traces (with and without noise). To provide some insights on the output returned by the approach, we discuss some of the partially ordered subgraphs obtained in the experiments without random noise along with the corresponding subgraphs. Note that edges in the subgraphs are associated with a label of the form headActivity_tailActivity, where headActvity represents the source node of the edge and tailActivity represents its target node.
Figure 15 shows two partially ordered subgraphs obtained using p-traces, namely PO17 and PO24. Partially ordered subgraph PO17 (Fig. 15a) represents one of the deviations that were inserted in the manipulated log, namely the swap of activities SRPP and FRPP. It consists of three anomalous subgraphs, i.e. SUB2 (Fig. 15c), SUB63 (Fig. 15d) and SUB105 (Fig. 15e). The ordering relations shown in the partially ordered subgraph can be easily justified taking into account the process model and the inserted deviations. In particular, SUB2 and SUB105 share one node, i.e. FRPP, which motivates the interleaving relation; SUB63 starts immediately after SUB2, therefore they are in a strictly sequential relation; finally, between SUB105 and SUB63 there is an eventually relation, since SUB63 occurs after SUB105 but not immediately after. Partially ordered subgraph PO24 (Fig. 15b) captures the replacement of activity SHRPP with the sequence of activities 〈SRP, FRP〉. It consists of two anomalous subgraphs, i.e. SUB2 (discussed above) and SUB17 (Fig. 15f), connected by a strictly sequential relation. These partially ordered subgraphs provide two significant examples of the capability of the approach in deriving anomalous subgraphs and in correlating them, thus providing a valuable support for the investigation of recurrent anomalous behaviors.
Although we were able to detect all the inserted deviations also using sequential traces, analysis of the partially ordered subgraphs shows that, as expected, several partially ordered subgraphs correspond to different sequential instantiations of the same parallel behavior. This significantly affects the usefulness of the obtained diagnostic information. First, it poses some challenges for the human analyst in interpreting the results. Consider, for instance, the partially ordered subgraphs in Fig. 16. These partially ordered subgraphs consist of only one anomalous subgraph representing a possible sequential instantiation of the swapping of activities SRPP and FRPP. It is straightforward to observe that this outcome is much less expressive and intuitive than the one obtained using p-traces and does not provide any insight regarding the actual control flow. Indeed, the analyst has to manually explore and compare the subgraphs to derive that they actually represent the same behavior. This is far from trivial even in this simple setting and when addressing complex, real-life processes can easily become a time-consuming, error prone task.
Another relevant drawback of neglecting concurrency is that it can easily lead to loose relevant information. The support of each sequential instantiation corresponds to a fraction of the support of the concurrent behavior. For instance, the support of the behavior showing the swapping of activities SRPP and FRPP is 56.4%; while the support of the subgraphs in Fig. 16 ranges from 9.5% to 12%. This can lead to miss some or all the instantiations of a parallel behavior, especially when its support is close to the threshold set for mining the itemsets. Furthermore, it makes it more challenging to detect correlated anomalous behaviors. A clear example of this issue is provided by the experiments without random noise on sequential traces; as shown in Table 2, all the partially ordered subgraphs mined in these experiments involve just one subgraph. Therefore, no correlation among deviations was detected; while in the experiment without random noise using p-traces, we were able to find several interesting correlations.
Finally, we would like to point out that, although the outcome obtained using sequential traces in presence of random noise looks simpler (and, hence, preferable for a human analyst) than the outcome obtained using p-traces, since it involves much less patterns, a deeper look to the obtained results highlighted the same drawbacks we discussed for the previous set of experiments, which are made even worse by the presence of random noise. In fact, although we were able to obtain some correlations, actually these correlations involve subgraphs representing portions of the same deviation, which explains why they frequently occur together. It is straightforward to see that these correlations are trivial and, thus, do not provide useful insights. Moreover, it is worth noting that in this experiment we also miss some of the sequential instantiation of parallel behaviors, which prevents the correct reconstruction of these behaviors. Although the presence of noise also affected partially ordered subgraphs mined from p-traces, leading to deviations spread within several subgraphs, we were anyway able to detect correlations involving different deviations; moreover, the partially ordered subgraphs obtained from p-traces properly reconstructed concurrent behaviors.

7.2 Real-life dataset: BPI 2012 challenge

Settings
To evaluate the applicability of our approach to real-life settings, we used the event log recording the loan management process of a Dutch Financial Institute, which was made available for the 2012 BPI challenge (2012). The event log contains the events recorded for three intertwined subprocesses: subprocess A specifies how loan applications should be handled, subprocess O describes how loan offers should be handled, and subprocess W specifies how work items are processed.
To reduce the overall complexity of the event log and improve the results of process discovery, we preprocessed the event log based on the findings of other researchers (Adriansyah and Buijs 2012; Bose and van der Aalst 2012). In particular, artificial start and end events were added to process instances, only events representing the completion of an activity were considered in the analysis, and potential redundant events were removed from the event log (Bose and van der Aalst 2012). To ensure that process loans are completed, we filtered out all process instances staring in 2012 (Adriansyah and Buijs 2012). After preprocessing, the event log contains 85,426 events in 7,455 event traces. We used inductive miner (Leemans et al. 2013) to discover a process model for this event log, shown in Fig. 17.
The log provided for the BPI challenge 2012 consists of sequential traces. We used Building Instance Graphs (BIG) algorithm (Diamantini et al. 2016) to construct p-traces from the sequential traces recorded in the event log. BIG algorithm exploits information encoded in a process model (i.e., causal relations between process activities) to construct p-traces. We applied our approach to the obtained process model and p-traces. Two different support thresholds were used to find frequent itemsets (i.e., 5% and 3%) and a threshold of 50% was used to derive ordering relations between subgraphs in a frequent itemset.
Results
Table 3 shows the results of our experiments. In total, 292 subgraphs and 96 anomalous subgraphs were obtained. Using a threshold of 5% for frequent itemsets, 36 partially ordered subgraphs with average support of 6.15% were obtained. By decreasing this threshold to 3%, 81 partially ordered subgraphs with average support of 3.49% were obtained. To provide a concrete example of the outcome of the approach and illustrate its capability, next we discuss in detail one of the discovered partially ordered subgraphs of each experiment, namely PO1 obtained from Exp1 and PO2 obtained from Exp2.
Table 3
Results of experiments on BPI Challenge 2012 event log
Experiment
Experiment setting
Subgraphs
Anomalous
P.O.
Average
Average
Average
(Ordering relation threshold,
subgraphs
support
subgraphs
activities
Frequent itemset threshold)
P.O.
per P.O.
per P. O.
BPI2012-Exp1
(50%, 5%)
292
96
36
6.15%
1.13
5.52
BPI2012-Exp2
(50%, 3%)
  
81
3.49%
1.56
7.5
Figure 18 shows PO1 along with its subgraphs. As shown in the figure, PO1 consists of two anomalous subgraphs, SUB28 and SUB174, connected through an interleaving relation. This relation is due to the fact that these two subgraphs share the same instance of activity O-Sent and, thus, they overlap. Subgraph SUB174 shows that, after the offer was sent (O-Sent), activity A-Canceled was skipped. In addition, subgraph SUB174 shows that, after filling in information for the application (W-Completerenaanvraag), another offer was sent to the client, which is not allowed by the process model in Fig. 17. Subgraph SUB28 shows that, after sending the offer, the client was contacted for obtaining more information. However, according to the process model in Fig. 17, activity W-Completerenaanvraag should be executed before contacting the client. Looking at the behavior represented by PO1 as a whole, we can observed that the offer (or a new offer) has been resent to the client after the information has been filled in for the application (W-Completerenaanvraag) and, then the client is contacted as prescribed by the model. We argue that further investigation should be performed to understand why multiple offers were sent to clients as well as why activity A-Canceled was skipped. It worth noting that some of the obtained partially ordered subgraphs (not reported here) also show that activity A-Canceled was often swapped with activity W-Completerenaanvraag.
It is worth mentioning that we did not discovered any partially ordered subgraph exhibiting parallel behavior with the setting of Exp1 (i.e., a threshold of 5% for frequent itemset and a threshold of 50% for ordering relations). This is due to the fact that concurrent behaviors have a very low support in the BPI 2012 challenge event log. By decreasing the threshold used for frequent itemset to 3%, we were able to obtain partially ordered subgraphs containing these subgraphs. Figure 19a shows one example of such partially ordered subgraphs. In particular, PO2 consists of one anomalous subgraph, namely SUB30 (Fig. 19b). This subgraph shows that the activities for assessing (W-Validerenaanvraag) and approving (A-Approved) the application were executed immediately after calling the client (W-Nabellenoffertes). In addition, it shows that, during the assessment phase (W-Validerenaanvraag) for approving the request (A-Approved), the client was not contacted for obtaining missing information (W-Nabellenincompletedossiers). These cases should be investigated and, if they are recognized as normal behavior, the process model should be repaired to reflect this behavior.
The results confirm that the patterns obtained from the proposed approach can highlight frequent and correlated anomalous behaviors. By analyzing the obtained patterns, the process model might be repaired to represent the actual expected behavior. For example, the process model can be extended in order to allow skipping of activity A-Canceled or swapping it with activity W-Completerenaanvraag. The analyst may also decide to investigate anomalous behaviors to ensure that they conform to the defined guidelines. For example, sending multiple offers (O-Sent) to a client should be investigated to ensure that it is a justified exception. If this behavior corresponds to errors, actions should be taken to prevent future occurrence of them.

7.3 Real-life dataset: BPI 2017 challenge

Settings
We also evaluated our approach using the event log that was made available for the 2017 BPI challenge (2017). This log records the executions of a process for handling credit requests within a financial company. The event log contains detailed information about applications submitted by clients, loan offers sent by the company, and work items processed by employees or by the system. At the end of the process, the submitted applications can be approved or declined by the company, or be canceled by the client.
To obtain a reliable process model, we preprocessed the event log. In particular, artificial start and end events were added to process instances. As it appears that process instances starting after October 2016 are not likely to be completed, we filtered them out. To avoid obtaining a spaghetti-like process model, we partition the event log into homogeneous subsets of process instances. In the experiments, we only focus on the analysis of applications that were denied. After preprocessing, the event log contains 124,866 events in 3,093 event traces. We used inductive miner (Leemans et al. 2013) to discover a process model for this event log; the discovered process is reported in Fig. 20.
The log provided for the BPI challenge 2017 consists of sequential traces. As we did for the 2012 BPI challenge log, we used the BIG algorithm to construct p-traces from the sequential traces recorded in the event log. We applied our approach to the obtained process model and p-traces using the same settings as described in Section 7.2.
Results
Table 4 shows the results of our experiments. In total, 670 subgraphs and 39 anomalous subgraphs were obtained. Using a threshold of 5% for frequent itemsets, 324 partially ordered subgraphs with average support of 8.02% were obtained. We also performed an experiment in which the threshold for frequent itemsets was set to 3%. In this case, we obtained 517 partially ordered subgraphs with average support of 6.23%. In these experiments, the constructed anomalous subgraphs are highly correlated. Thus, in comparison to the results presented in Section 7.2, the obtained partially ordered subgraphs on average contain a larger number of anomalous subgraphs.
Table 4
Results of experiments on BPI Challenge 2017 event log
Experiment
Experiment setting
Subgraphs
Anomalous
P.O.
Average
Average
Average
(Ordering relation threshold,
subgraphs
support
subgraphs
activities
Frequent itemset threshold)
P.O.
per P.O.
per P. O.
BPI2017-Exp1
(50%, 5%)
670
39
324
8.02%
3.40
21.48
BPI2017-Exp2
(50%, 3%)
  
517
6.23%
3.47
22.70
Differently from the experiments on the BPI 2012 challenge log, we discovered a number of partially ordered subgraphs exhibiting concurrency when the threshold for frequent itemsets was set to 5%. Next, we discuss in detail two of these partially ordered subgraphs, namely PO32 and PO140, to provide a concrete example of the outcome of the approach and illustrate its capabilities.
Figure 21 shows PO32 along with its subgraphs, namely SUB3, SUB65 and SUB92. SUB3 is connected to SUB65 and SUB92 through eventually relations and SUB92 is connected to SUB65 through a sequential relation. SUB3 shows that W-Complete-application-suspend was skipped, which is not allowed by the process model. In general, when an activity is started, it might be suspended (temporarily halted) and resumed again but these steps in the life-cycle of activities are not always necessary. SUB92 shows that W-Validate-application-schedule and O-Returned can be executed together and offers might be returned by clients while applications are being validated. However, according to the model, either W-Validate-application-schedule or O-Returned should be executed and the application should be validated only after an offer is returned. According to the transnational life-cycle model (van der Aalst 2016), an activity should be scheduled before it can start. Thus, W-Validate-application-schedule should always be executed before the execution of W-Validate-application-start. Moreover, we observed that in 2789 cases O-Returned was executed immediately after the execution of activities W-Validate-application-start and A-Validating. This means that the preliminary validation of an application (e.g., checking the submitted documents) started before an offer is returned, indicating a swapping between these two activities. SUB65 shows that some applications were denied (A-Denied) and offers were refused (O-Refused) immediately after the validation of applications is resumed (W-Validate-application-resume) or suspended (W-validate-application-suspended), which is not allowed by the process model. An application can be denied at different states of a process execution if it does not fit acceptance criteria. For example, along with an offer other documents such as payslips and bank statements might be required to be provided by a client. If the income is not sufficient, the application is denied even though the client returned the offer. This indicates that applications could be potentially denied also during the validation phase. Partially ordered subgraph PO32 shows that these behaviors can frequently occur together. This larger view of the deviant behavior shows that after an offer is returned the validation process can be suspended and resumed, and the application denied and the corresponding offer refused. As shown in the process model of Fig. 20, these activities belong to different branches of a choice construct. This suggests that the validation process and refusal of offers can be reiterated as represented by the (red) loop in Fig. 20.
Figure 22 shows PO140 along with two of its subgraphs, namely SUB4 and SUB15. SUB92 is connected to SUB4 through a sequentially relation and SUB92 and SUB4 are connected to SUB15 through eventually relations. As discussed before, SUB92 (see Fig. 21c) indicates that offers might be returned by clients while applications are being validated. SUB4 shows that the validation phase can be suspended and resumed again, while the process model does not allow it. According to the standard transnational life-cycle model (van der Aalst 2016), this behavior should be allowed. Thus, the process model should be revised to reflect this finding. SUB15 shows that O-Refused was performed multiple times, while according to the process model it must be performed once. During the handling of a credit request, multiple offers might be sent to the client. In case the application is denied, all active offers should be refused.
The results confirm that the patterns obtained from the proposed approach can highlight frequent and correlated anomalous behaviors. In particular, patterns provide more accurate diagnostics and a better understanding of deviant behaviors. Based on the findings obtained from the analysis of the mined patterns, the process model might be repaired to better reflect the reality. For example, the process model can be extended in order to allow repetition of O-Refused, swapping of O-Returned with W-Validate-application-start and A-Validating, and the execution of A-Denied and O-Refused during the validation phase.
This work embraces two main research areas, namely subgraph extraction and conformance checking. In this section, we discuss recent developments in these areas.
Subgraph extraction
Several approaches for subgraph extraction have been proposed in the area of business processes. Bose and van der Aalst (2009) detect subprocesses by identifying sequences of events that fit a-priori defined templates; Huang et al. (2012) exploit a sequence pattern mining algorithm to derive frequent sequences of clinical activities from clinical logs; Leemans and van der Aalst (2014) introduce an approach to derive “episodes”, i.e. directed graphs where nodes correspond to activities and edges to eventually-follow precedence relations, which, given a pair of activities, state which one occurs later. Compared to these approaches, our work does not require defining any predefined template and extracts the subprocesses that are the most relevant according to their description length, thus taking into account both frequency and size in determining the relevance of subprocesses.
Other approaches aim to convert traces into directed graphs representing execution flows and, then, apply frequent subgraph mining techniques to derive the most relevant subgraphs. For instance, Hwang et al. (2004) generate “temporal graphs”, where two nodes are linked only if the corresponding activities have to be executed sequentially. The applicability of this approach, however, is limited to event logs storing starting and completion time of events. Greco et al. (2005) propose an FSM algorithm that exploits knowledge about relationships among activities (e.g., AND/OR split) to drive subgraphs mining. Graphs are generated by replaying traces over the process model; however, this algorithm requires a model properly representing the event log, which may not be available for many real-world processes. In contrast, our approach for subprocess extraction does not require neither the presence of special attributes in the event log nor a-priori models of the process or other domain knowledge. Moreover, Greco and colleagues exploit the relations shown in the process model to speed up the candidates generation step in the subgraph mining algorithm. However, this strategy is not suited for our purposes and might actually hinder the extraction of subgraphs involving deviations, since these subgraphs are not shown in the model. Greco and colleagues have extended their approach in Greco et al. (2007) to mine both connected and unconnected subgraphs from workflow executions. They implement an a-priori algorithm to determine, among the set of frequent connected subgraphs mined with the approach proposed in Greco et al. (2005), the ones that co-occur with a frequency above a given threshold. However, the authors do not explore ordering relations among the set of co-occurring subgraphs. Moreover, also in this case, they exploit knowledge from the model to speed up the search, which, as mentioned before, is not an effective strategy when dealing with anomalous subgraphs.
It is worth noting that once traces have been converted into directed graphs, one can apply any frequent subgraph mining technique to derive the subgraphs of interest. Subgraph mining is an active research area; a plethora of FSM approaches have been proposed in literature during the last years (see, e.g., Jiang et al. 2013 for a survey). Most of well-known techniques are complete approaches (e.g., AGM (Inokuchi et al. 2000) and FSG (Kuramochi and Karypis 2001) algorithms), namely they aim at deriving the entire set of subgraphs that fit user-defined requirements (typically, minimum support). Nevertheless, some heuristics techniques have been developed as well, like the SUBDUE algorithm, used in this work, or the GREW algorithm (Kuramochi and Karypis 2004). Heuristic techniques sacrifice completeness for efficiency. Namely, they do not aim to return the complete set of subgraphs, but only of those considered the most relevant according to a given metric. It is worth noting, however, that to the best of our knowledge, only SUBDUE exploits the description length metric to infer the most relevant subgraphs. Moreover, SUBDUE also supports the exploration of inclusion relationships between subgraphs by returning a hierarchy in which subgraphs are arranged according to these relationships. The use of taxonomies for the analysis of process behaviors has also been proposed by other process discovery techniques that aim to extract from a set of process executions a process model able to represent the underlying process (van der Aalst et al. 2003). In this context, taxonomies are used to enable an exploration of the process on different levels of abstractions. For instance, Bose et al. (2011) propose to replace the subprocesses inferred with the technique proposed in Bose and van der Aalst (2009) with single activities and, then, to mine a process model in which most frequent subprocesses are represented by a single node. This procedure can be repeated until the desired level of abstraction is reached. At the end, a taxonomy of process models is obtained, where the model at each level represents the process with a degree of abstraction higher than the models at lower levels. Mannhardt and Tax (2017) adopt a similar idea, using local process models discovered by the technique proposed in Tax et al. (2016) to determine the set of activities to replace. A different approach is proposed in Greco et al. (2008), which iteratively refines the models at the current level of the taxonomy by properly clustering traces that fit such models and, then, derives from the clusters new models representing a portion of the behavior of the parent model with a higher level of detail than the latter. The taxonomies obtained using the aforementioned approaches differ from the one obtained using SUBDUE, which have been exploited in our work, both by construction and purpose. Indeed, our approach exploits SUBDUE hierarchy simply as a means to derive subgraphs inclusion relations, rather than to provide users with a multi-level process exploration tool.
Conformance checking
Conformance checking aims to verify whether the observed behavior recorded in an event log matches the intended behavior represented as a process model. Several efforts have been devoted in the last years to checking conformance and a number of techniques have been proposed to verify the conformity of event logs with process specifications. Some approaches (Borrego and Barba 2014; Caron et al. 2013; Taghiabadi et al. 2014) propose to check whether event traces satisfy a set of compliance rules. Rozinat and van der Aalst (2008) propose a token-based technique to replay event traces over a process model and use the information obtained from remaining and missing tokens to detect deviations. However, it has been shown that token-based techniques can provide misleading diagnostics. Recently, alignments have been proposed as a robust approach to conformance checking (van der Aalst et al. 2012). Alignments are able to pinpoint deviations causing nonconformity based on a given cost function. These cost functions, however, are usually based on human judgment and, hence, prone to imperfections, which can ultimately lead to incorrect diagnostics. To obtain probable explanations of nonconformity, Alizadeh et al. (2014) propose an approach to compute the cost function by analyzing historical logging data, which is extended in Alizadeh et al. (2015) to consider multiple process perspectives. However, these techniques are only able to deal with sequential event traces (i.e., total ordered traces); thus, diagnostics can be unreliable when timestamps of events are coarse or incorrect.
A number of graph-based approaches for anomaly detection have been proposed to deal with concurrency in process executions (see Akoglu et al. 2015 for a survey). For instance, Eberle et al. (2010) present graph-based anomaly detection (GBAD). GBAD comprises three anomaly detection algorithms, each of them tailored to discover a certain type of anomaly, namely insertions, modifications and deletions. Similarly to our work, GBAD relies on SUBDUE to discover frequent subgraphs in a given graph; however, in GBAD the identified subgraphs are treated as normative patterns and are used as the baseline for the detection of anomalies. In contrast, our approach uses the subgraphs mined using SUBDUE as a representation of the frequent behaviors observed by the system whose compliance is assessed against a process model. Lu et al. (2016) propose an approach to detect deviations in an event log by identifying frequent common behavior and uncommon behavior. In particular, this work computes mappings between events, thus highlighting similar and dissimilar behavior between process instances. Similarly to graph-based approaches, the work in Lu et al. (2016) assumes that deviations are uncommon behavior; however, it is limited to deviations corresponding to insertion, i.e. activities that have been executed but are not allowed in the normative process. Process instances are fused in a representative execution graph based on the similarity of events. This graph, however, can provide misleading diagnostics about deviations as it is not able to properly represent correlated anomalous behaviors. In contrast, our approach builds anomalous patterns that account for behaviors that frequently happen together.
Most of the existing techniques for conformance checking, as the ones discussed above, are only able to detect simple anomalous behaviors like insertions, modifications and deletions of activities. Only a few conformance checking techniques attempt to discover more complex anomalies. For instance, Banescu et al. (2012) use a token-based approach for checking conformance of an event log and a process model and classify anomalies by analyzing the configuration of missing and added tokens using deviation patterns. Adriansyah et al. (2013) define anomalous patterns representing swappings and replacements of activities. These patterns are combined with the process model, and alignment-based techniques are used to discover their occurrence in the log. However, these approaches are only able to detect complex anomalies based on predefined anomalous patterns. In contrast, the approach proposed in this work does not rely on predefined anomalous patterns. Rather, our work aims to extract anomalous patterns representing frequent anomalous behavior from historical logging data.
In general, conformance checking techniques, including graph-based approaches for anomaly detection, differ from our approach in that they usually aim to detect particular instances of anomalous behavior whereas our approach aim to discover and analyze recurrent anomalous behaviors of arbitrary complexity. To the best of our knowledge, this is the first work that provides a systematic solution for the discovery and analysis of recurrent anomalous behaviors in the context of business process compliance.
It is worth noting that recently decomposition approaches have been introduced to reduce the computational complexity of conformance checking techniques (de Leoni et al. 2014; van der Aalst 2013). These techniques propose to split a process model (typically represented as a Petri net) in a set of subprocesses from which it is possible to reconstruct the original model. The event log is then split in a set of sublogs, each of them obtained by projecting the original log on the set of activities related to one of the subprocesses. Conformance checking is applied to pairs of sublog and subprocesses, which allows a significant reduction of the computational complexity. Although in principle this approach might be applied in combination of our approach to speed up the extraction of subgraphs and the conformance checking step, it likely leads to some information loss. In fact, if subgraphs are extracted from sublogs, subgraphs involving activities belonging to different subprocesses cannot be captured.

9 Conclusions and future work

In this work, we have presented an extension to our previous approach for the discovery of anomalous patterns from historical logging data. In particular, the obtained patterns represent recurrent high-level deviations in process executions, which can be spanned across different (and not connected) portions of the process. The main novelty of the extended approach consists in the capability of dealing with concurrency, thus providing patterns that better reflects the control flow of processes. In particular, taking into account possible parallelisms enables a more accurate diagnosis of anomalous behaviors and provides analysts with a more comprehensive and compact representation of deviations.
The extended approach extracts relevant subgraphs from partially ordered traces, explicitly modeling possible parallelisms, rather than from totally ordered traces that are typically used by classic conformance checking techniques. This allows us to obtain relevant subgraphs that better reflect the control flow of processes. To identify anomalous subgraphs, we have proposed a novel conformance checking algorithm tailored to check the conformance of partially ordered subtraces exhibiting concurrent behavior. Moreover, we have investigated and formalized ordering relations between subgraphs exhibiting concurrent behavior with respect to a process model. Based on the identified relations, we have shown how to infer partially ordered subgraphs exhibiting correlations among recurrent anomalous behaviors. Finally, we have developed a set of guidelines to transform partially ordered subgraphs into anomalous patterns expressed as Petri nets.
The approach has been implemented as a plug-in of the Esub tool and has been validated using both synthetic and real-life logs. The experiments demonstrated the capability of the approach to return meaningful patterns capturing high-level deviations that, on the other hand, would have been difficult to identify without accounting for concurrency.
The anomalous patterns obtained using our approach can be exploited for several purposes. Among possible applications, we are investigating their application for online monitoring. This would allow for early detection of occurrences of recurring anomalous behaviors, for which accurate diagnostics are already available. Moreover, we are investigating how anomalous patterns can be exploited to guide the definition of measures for preventing and/or responding to anomalous behaviors, especially in the security context. In this work, we have provided an approach for the design of anomalous patterns from partially ordered subgraphs. However, additional support should be provided to analysts for the design of anomalous patterns. To this end, we are investigating approaches for the (semi)automated generation of anomalous patterns from partially ordered subgraphs. We also plan to investigate the use of decomposing techniques (de Leoni et al. 2014; van der Aalst 2013) in combination to our approach. In particular, investigation is required to understand to what extent these techniques affect the mining of anomalous patterns and the trade-off between information loss and efficiency gain. Finally, in our study we observed that logs often record process executions as sequential traces of events. Although there exist several techniques for the generation of p-traces from sequential traces, the adopted technique can influence the (anomalous) behaviors that can be extracted using our approach and, thus, the patterns that can be obtained. An interesting direction for future work is the study of these techniques and, in particular, their impact on the quality of the patterns extracted using the proposed approach.

Acknowledgments

This work has been partially funded by the NWO CyberSecurity programme under the PriCE project and by the ITEA2 project M2MGrid (No. 13011).
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
Appendix

Appendix A: Converting a p-trace into a Petri net

Algorithm 4 describes the conversion from a p-trace to a Petri net. The set of places is initialized by creating an initial place pini and an ending place pend (line 1). The set of transitions in the Petri net consists of the nodes in the p-trace (line 2). We also create two invisible transitions, invini and invend, which are connected to pini and pend respectively (line 3). Invisible transition invini is needed to capture the fact that events in Eini, i.e. the set of transitions for which the corresponding node in the p-trace does not have an incoming edge, can occur concurrently. Similarly, invend is needed to capture the fact that events in Eend, i.e. the set of transitions for which the corresponding node in the p-trace does not have an outgoing edge, can occur concurrently. Note that invini (invend respectively) can be omitted if Eini (Eend respectively) consists of only one transition.
Then, a place p is added for each edge in the p-trace. Transitions and places are connected in such a way that for every (e, e) ∈ W we have (e, p), (p, e) ∈ F. Each transition ei for which the corresponding node in the p-trace does not have an incoming edge (i.e., eiEini) is connected to invisible transition invini through a newly created place pi (lines 9–11). Similarly, each transition ej for which the corresponding node in the p-trace does not have an outgoing edge (i.e., ejEend) is connected to invisible transition invend through a newly created place pj (lines 12–14). The initial marking mi and final marking mf of the Petri net consist of pini and pend respectively, i.e. mi = [pini] and mf = [pend].
Footnotes
1
\(\mathbb {B}(X)\) represents the set of all multisets over X.
 
2
Given a set Γ,a partial order over Γis a binary relation ≼⊆Γ×Γthat is i) irreflexive, i.e. \(\gamma \npreceq \gamma \), ii) antisymmetric, i.e. if γγ,then \(\gamma ^{\prime } \npreceq \gamma \)and iii) transitive, i.e. if γγand γγ,then γγ.
 
3
The complete money transfer process can be found in van den Broucke et al. (2014).
 
4
The definition of relevance depends on the FSM algorithm exploited to derive the subgraphs. In this work, we have adopted DL (see Section 4).
 
5
The term “ordered pair” should not be confused with ordering relation. Here, ordered pair indicates that we assess the ordering relation between the first element and the second element of the ordered pair.
 
6
Recall from Section 2 that a marking is reachable by itself with a firing sequence of length 0.
 
7
The Petri net representation of a subgraph can be obtained similarly to the one of a p-trace using Algorithm 4 in Appendix. For the sake of convenience, in Fig. 11 we represent the initial and final place of (the Petri net representation of) a subgraph externally to the rectangle representing the subgraph. For instance, in Fig. 11a, p1 is the initial place of \(\overline {{\gamma }_{i}}\), p2 is the final place of \(\overline {{\gamma }_{i}}\) and the initial place of \(\overline {{\gamma }_{j}}\), and p3 is the final place of \(\overline {{\gamma }_{j}}\).
 
8
In the anomalous pattern of Fig. 13, invisible transition Inv1 is introduced by Algorithm 4 (Appendix) when transforming subgraph γ12 into a Petri net to capture the concurrency between activities RBPC, RIBPC and REPC.
 
9
Note that the traditional SUBDUE implementation takes a single graph as input. Therefore, we generated a graph consisting of several disjoint (sub)graphs, each corresponding to a p-trace of the event log. The resulting graph was given as input to SUBDUE.
 
Literature
go back to reference Adriansyah, A., & Buijs, J.M. (2012). Mining process performance from event logs: the BPI challenge 2012 case study. BPM center report BPM-12-15 BPMcenter.org. Adriansyah, A., & Buijs, J.M. (2012). Mining process performance from event logs: the BPI challenge 2012 case study. BPM center report BPM-12-15 BPMcenter.org.
go back to reference Adriansyah, A., van Dongen, B.F., & van der Aalst, W.M.P. (2011). Conformance checking using cost-based fitness analysis. In Proceedings of IEEE international enterprise distributed object computing conference (pp. 55–64). Piscataway: IEEE. Adriansyah, A., van Dongen, B.F., & van der Aalst, W.M.P. (2011). Conformance checking using cost-based fitness analysis. In Proceedings of IEEE international enterprise distributed object computing conference (pp. 55–64). Piscataway: IEEE.
go back to reference Adriansyah, A., van Dongen, B.F., & Zannone, N. (2013). Controlling break-the-glass through alignment. In Proceedings of international conference on social computing (pp. 606–611). Piscataway: IEEE. Adriansyah, A., van Dongen, B.F., & Zannone, N. (2013). Controlling break-the-glass through alignment. In Proceedings of international conference on social computing (pp. 606–611). Piscataway: IEEE.
go back to reference Akoglu, L., Tong, H., & Koutra, D. (2015). Graph based anomaly detection and description: a survey. Data Mining and Knowledge Discovery, 29(3), 626–688.MathSciNetCrossRef Akoglu, L., Tong, H., & Koutra, D. (2015). Graph based anomaly detection and description: a survey. Data Mining and Knowledge Discovery, 29(3), 626–688.MathSciNetCrossRef
go back to reference Alizadeh, M., de Leoni, M., & Zannone, N. (2014). History-based construction of alignments for conformance checking: Formalization and implementation. In Data-driven process discovery and analysis, LNBIP 237 (pp. 58–78). Berlin: Springer. Alizadeh, M., de Leoni, M., & Zannone, N. (2014). History-based construction of alignments for conformance checking: Formalization and implementation. In Data-driven process discovery and analysis, LNBIP 237 (pp. 58–78). Berlin: Springer.
go back to reference Alizadeh, M., de Leoni, M., & Zannone, N. (2015). Constructing probable explanations of nonconformity: a data-aware and history-based approach. In Proceedings of symposium series on computational intelligence (pp. 1358–1365). Piscataway: IEEE. Alizadeh, M., de Leoni, M., & Zannone, N. (2015). Constructing probable explanations of nonconformity: a data-aware and history-based approach. In Proceedings of symposium series on computational intelligence (pp. 1358–1365). Piscataway: IEEE.
go back to reference Banescu, S., Petkovic, M., & Zannone, N. (2012). Measuring privacy compliance using fitness metrics. In Business process management, LNCS 7481 (pp. 114–119). Berlin: Springer. Banescu, S., Petkovic, M., & Zannone, N. (2012). Measuring privacy compliance using fitness metrics. In Business process management, LNCS 7481 (pp. 114–119). Berlin: Springer.
go back to reference Borrego, D., & Barba, I. (2014). Conformance checking and diagnosis for declarative business process models in data-aware scenarios. Expert Systems with Applications, 41 (11), 5340–5352.CrossRef Borrego, D., & Barba, I. (2014). Conformance checking and diagnosis for declarative business process models in data-aware scenarios. Expert Systems with Applications, 41 (11), 5340–5352.CrossRef
go back to reference Bose, R.P.J.C., & van der Aalst, W.M.P. (2009). Abstractions in process mining: a taxonomy of patterns. In Business process management, LNCS 5701 (pp. 159–175). Berlin: Springer. Bose, R.P.J.C., & van der Aalst, W.M.P. (2009). Abstractions in process mining: a taxonomy of patterns. In Business process management, LNCS 5701 (pp. 159–175). Berlin: Springer.
go back to reference Bose, R.P.J.C., & van der Aalst, W.M.P. (2012). Process mining applied to the BPI challenge 2012: divide and conquer while discerning resources. In Business process management (pp. 221–222). Berlin: Springer. Bose, R.P.J.C., & van der Aalst, W.M.P. (2012). Process mining applied to the BPI challenge 2012: divide and conquer while discerning resources. In Business process management (pp. 221–222). Berlin: Springer.
go back to reference Bose, R.P.J.C., Verbeek, H.M.W., & van der Aalst, W.M.P. (2011). Discovering hierarchical process models using ProM. In CAISE forum, (Vol. 107 pp. 33–48). Berlin: Springer. Bose, R.P.J.C., Verbeek, H.M.W., & van der Aalst, W.M.P. (2011). Discovering hierarchical process models using ProM. In CAISE forum, (Vol. 107 pp. 33–48). Berlin: Springer.
go back to reference Caron, F., Vanthienen, J., & Baesens, B. (2013). Comprehensive rule-based compliance checking and risk management with process mining. Decision Support Systems, 54(3), 1357–1369.CrossRef Caron, F., Vanthienen, J., & Baesens, B. (2013). Comprehensive rule-based compliance checking and risk management with process mining. Decision Support Systems, 54(3), 1357–1369.CrossRef
go back to reference de Leoni, M., Munoz-Gama, J., Carmona, J., & van der Aalst, W.M.P. (2014). Decomposing conformance checking on Petri nets with data. BPM Report BPM-14-06 BPMcenter.org. de Leoni, M., Munoz-Gama, J., Carmona, J., & van der Aalst, W.M.P. (2014). Decomposing conformance checking on Petri nets with data. BPM Report BPM-14-06 BPMcenter.org.
go back to reference Diamantini, C., Genga, L., & Potena, D. (2015). ESub: exploration of subgraphs. In Proceedings of the BPM demo session. CEUR-WS.org (pp. 70–74). Diamantini, C., Genga, L., & Potena, D. (2015). ESub: exploration of subgraphs. In Proceedings of the BPM demo session. CEUR-WS.org (pp. 70–74).
go back to reference Diamantini, C., Genga, L., Potena, D., & van der Aalst, W.M.P. (2016). Building instance graphs for highly variable processes. Expert Systems with Applications, 59, 101–118.CrossRef Diamantini, C., Genga, L., Potena, D., & van der Aalst, W.M.P. (2016). Building instance graphs for highly variable processes. Expert Systems with Applications, 59, 101–118.CrossRef
go back to reference Eberle, W., Graves, J., & Holder, L. (2010). Insider threat detection using a graph-based approach. Journal of Applied Security Research, 6(1), 32–81.CrossRef Eberle, W., Graves, J., & Holder, L. (2010). Insider threat detection using a graph-based approach. Journal of Applied Security Research, 6(1), 32–81.CrossRef
go back to reference Fahland, D. (2008). Translating UML2 activity diagrams petri nets for analyzing IBM WebSphere business modeler process models. Informatik-berichte 226 Humboldt-Universität zu Berlin. Fahland, D. (2008). Translating UML2 activity diagrams petri nets for analyzing IBM WebSphere business modeler process models. Informatik-berichte 226 Humboldt-Universität zu Berlin.
go back to reference Fahland, D., & van der Aalst, W.M.P. (2015). Model repair—aligning process models to reality. Information Systems, 47, 220–243.CrossRef Fahland, D., & van der Aalst, W.M.P. (2015). Model repair—aligning process models to reality. Information Systems, 47, 220–243.CrossRef
go back to reference Genga, L., Alizadeh, M., Potena, D., Diamantini, C., & Zannone, N. (2017). APD tool: mining anomalous patterns from event logs. In Proceedings of the BPM demo track and BPM dissertation award co-located with 15th international conference on business process modeling, vol. 1920 of CEUR workshop proceedings. CEUR-WS.org. Genga, L., Alizadeh, M., Potena, D., Diamantini, C., & Zannone, N. (2017). APD tool: mining anomalous patterns from event logs. In Proceedings of the BPM demo track and BPM dissertation award co-located with 15th international conference on business process modeling, vol. 1920 of CEUR workshop proceedings. CEUR-WS.org.
go back to reference Genga, L., Potena, D., Martino, O., Alizadeh, M., Diamantini, C., & Zannone, N. (2016). Subgraph mining for anomalous pattern discovery in event logs. In Proceedings of international workshop on new frontiers in mining complex patterns. Berlin: Springer. Genga, L., Potena, D., Martino, O., Alizadeh, M., Diamantini, C., & Zannone, N. (2016). Subgraph mining for anomalous pattern discovery in event logs. In Proceedings of international workshop on new frontiers in mining complex patterns. Berlin: Springer.
go back to reference Greco, G., Guzzo, A., Manco, G., & Saccà, D. (2005). Mining and reasoning on workflows. IEEE Transactions on Knowledge and Data Engineering, 17(4), 519–534.CrossRefMATH Greco, G., Guzzo, A., Manco, G., & Saccà, D. (2005). Mining and reasoning on workflows. IEEE Transactions on Knowledge and Data Engineering, 17(4), 519–534.CrossRefMATH
go back to reference Greco, G., Guzzo, A., Manco, G., & Saccà, D. (2007). Mining unconnected patterns in workflows. Information Systems, 32(5), 685–712.CrossRef Greco, G., Guzzo, A., Manco, G., & Saccà, D. (2007). Mining unconnected patterns in workflows. Information Systems, 32(5), 685–712.CrossRef
go back to reference Greco, G., Guzzo, A., & Pontieri, L. (2008). Mining taxonomies of process models. Data & Knowledge Engineering, 67(1), 74–102.CrossRef Greco, G., Guzzo, A., & Pontieri, L. (2008). Mining taxonomies of process models. Data & Knowledge Engineering, 67(1), 74–102.CrossRef
go back to reference Han, J., Pei, J., & Yin, Y. (2000). Mining frequent patterns without candidate generation. SIGMOD Record, 29(2), 1–12.CrossRef Han, J., Pei, J., & Yin, Y. (2000). Mining frequent patterns without candidate generation. SIGMOD Record, 29(2), 1–12.CrossRef
go back to reference Holder, L., Cook, D., & Djoko, S. (1994). Substructure discovery in the SUBDUE system. In Proceedings of AAAI workshop on knowledge discovery in databases (pp. 169–180). Holder, L., Cook, D., & Djoko, S. (1994). Substructure discovery in the SUBDUE system. In Proceedings of AAAI workshop on knowledge discovery in databases (pp. 169–180).
go back to reference Huan, J., Wang, W., Prins, J., & Yang, J. (2004). Spin: mining maximal frequent subgraphs from graph databases. In Proceedings of SIGKDD international conference on knowledge discovery and data mining (pp. 581–586). New York: ACM. Huan, J., Wang, W., Prins, J., & Yang, J. (2004). Spin: mining maximal frequent subgraphs from graph databases. In Proceedings of SIGKDD international conference on knowledge discovery and data mining (pp. 581–586). New York: ACM.
go back to reference Huang, Z., Lu, X., & Duan, H. (2012). On mining clinical pathway patterns from medical behaviors. Artificial Intelligence in Medicine, 56(1), 35–50.CrossRef Huang, Z., Lu, X., & Duan, H. (2012). On mining clinical pathway patterns from medical behaviors. Artificial Intelligence in Medicine, 56(1), 35–50.CrossRef
go back to reference Hwang, S., Wei, C., & Yang, W. (2004). Discovery of temporal patterns from process instances. Computers in Industry, 53(3), 345–364.CrossRef Hwang, S., Wei, C., & Yang, W. (2004). Discovery of temporal patterns from process instances. Computers in Industry, 53(3), 345–364.CrossRef
go back to reference Inokuchi, A., Washio, T., & Motoda, H. (2000). An apriori-based algorithm for mining frequent substructures from graph data. In Proceedings of european conference on principles of data mining and knowledge discovery (pp. 13–23). Berlin: Springer. Inokuchi, A., Washio, T., & Motoda, H. (2000). An apriori-based algorithm for mining frequent substructures from graph data. In Proceedings of european conference on principles of data mining and knowledge discovery (pp. 13–23). Berlin: Springer.
go back to reference Jiang, C., Coenen, F., & Zito, M. (2013). A survey of frequent subgraph mining algorithms. The Knowledge Engineering Review, 28(01), 75–105.CrossRef Jiang, C., Coenen, F., & Zito, M. (2013). A survey of frequent subgraph mining algorithms. The Knowledge Engineering Review, 28(01), 75–105.CrossRef
go back to reference Jonyer, I., Cook, D., & Holder, L. (2002). Graph-based hierarchical conceptual clustering. Journal of Machine Learning Research, 2, 19–43.MATH Jonyer, I., Cook, D., & Holder, L. (2002). Graph-based hierarchical conceptual clustering. Journal of Machine Learning Research, 2, 19–43.MATH
go back to reference Kuramochi, M., & Karypis, G. (2001). Frequent subgraph discovery. In Proceedings IEEE international conference on data mining (pp. 313–320). Piscataway: IEEE. Kuramochi, M., & Karypis, G. (2001). Frequent subgraph discovery. In Proceedings IEEE international conference on data mining (pp. 313–320). Piscataway: IEEE.
go back to reference Kuramochi, M., & Karypis, G. (2004). GREW – A scalable frequent subgraph discovery algorithm. In Proceedings of IEEE international conference on data mining (pp. 439–442). Piscataway: IEEE. Kuramochi, M., & Karypis, G. (2004). GREW – A scalable frequent subgraph discovery algorithm. In Proceedings of IEEE international conference on data mining (pp. 439–442). Piscataway: IEEE.
go back to reference Lassen, K.B., & van Dongen, B.F. (2008). Translating message sequence charts to other process languages using process mining. In Transactions on petri nets and other models of concurrency I (pp. 71–85). Berlin: Springer. Lassen, K.B., & van Dongen, B.F. (2008). Translating message sequence charts to other process languages using process mining. In Transactions on petri nets and other models of concurrency I (pp. 71–85). Berlin: Springer.
go back to reference Leemans, M., & van der Aalst, W.M.P. (2014). Discovery of frequent episodes in event logs. In Proceedings of international symposium on data-driven process discovery and analysis. CEUR-ws.org (pp. 1–31). Leemans, M., & van der Aalst, W.M.P. (2014). Discovery of frequent episodes in event logs. In Proceedings of international symposium on data-driven process discovery and analysis. CEUR-ws.org (pp. 1–31).
go back to reference Leemans, S.J., Fahland, D., & van der Aalst, W.M.P. (2013). Discovering block-structured process models from event logs - a constructive approach. In Applications and theory of petri nets and concurrency (pp. 311–329). Berlin: Springer. Leemans, S.J., Fahland, D., & van der Aalst, W.M.P. (2013). Discovering block-structured process models from event logs - a constructive approach. In Applications and theory of petri nets and concurrency (pp. 311–329). Berlin: Springer.
go back to reference Lu, X., Fahland, D., & van der Aalst, W.M.P. (2014). Conformance checking based on partially ordered event data. In Business process management (pp. 75–88). Berlin: Springer. Lu, X., Fahland, D., & van der Aalst, W.M.P. (2014). Conformance checking based on partially ordered event data. In Business process management (pp. 75–88). Berlin: Springer.
go back to reference Lu, X., Fahland, D., van den Biggelaar, F.J.H.M., & van der Aalst, W.M.P. (2016). Detecting deviating behaviors without models. In Business process management workshops (pp. 126–139). Berlin: Springer. Lu, X., Fahland, D., van den Biggelaar, F.J.H.M., & van der Aalst, W.M.P. (2016). Detecting deviating behaviors without models. In Business process management workshops (pp. 126–139). Berlin: Springer.
go back to reference Mannhardt, F., & Tax, N. (2017). Unsupervised event abstraction using pattern abstraction and local process models. arXiv:1704.03520. Mannhardt, F., & Tax, N. (2017). Unsupervised event abstraction using pattern abstraction and local process models. arXiv:1704.​03520.
go back to reference Messmer, B.T., & Bunke, H. (1998). A new algorithm for error-tolerant subgraph isomorphism detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(5), 493–504.CrossRef Messmer, B.T., & Bunke, H. (1998). A new algorithm for error-tolerant subgraph isomorphism detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(5), 493–504.CrossRef
go back to reference Ouyang, C., Dumas, M., van der Aalst, W.M.P., Hofstede, A., & Mendling, J. (2009). From business process models to process-oriented software systems. ACM Transactions on Software Engineering and Methodology, 19(1), 2:1–2:37.CrossRef Ouyang, C., Dumas, M., van der Aalst, W.M.P., Hofstede, A., & Mendling, J. (2009). From business process models to process-oriented software systems. ACM Transactions on Software Engineering and Methodology, 19(1), 2:1–2:37.CrossRef
go back to reference Ramezani, E., Fahland, D., & van der Aalst, W.M.P. (2012). Where did I misbehave? Diagnostic information in compliance checking. In Business process management (pp. 262–278). Berlin: Springer. Ramezani, E., Fahland, D., & van der Aalst, W.M.P. (2012). Where did I misbehave? Diagnostic information in compliance checking. In Business process management (pp. 262–278). Berlin: Springer.
go back to reference Rozinat, A., & van der Aalst, W.M.P. (2008). Conformance checking of processes based on monitoring real behavior. Information Systems, 33(1), 64–95.CrossRef Rozinat, A., & van der Aalst, W.M.P. (2008). Conformance checking of processes based on monitoring real behavior. Information Systems, 33(1), 64–95.CrossRef
go back to reference Schmidt, D.C., & Druffel, L.E. (1976). A fast backtracking algorithm to test directed graphs for isomorphism using distance matrices. Journal of the ACM, 23(3), 433–445.MathSciNetCrossRefMATH Schmidt, D.C., & Druffel, L.E. (1976). A fast backtracking algorithm to test directed graphs for isomorphism using distance matrices. Journal of the ACM, 23(3), 433–445.MathSciNetCrossRefMATH
go back to reference Taghiabadi, E.R., Gromov, V., Fahland, D., & van der Aalst, W.M.P. (2014). Compliance checking of data-aware and resource-aware compliance requirements. In On the move to meaningful internet systems, LNCS 8841 (pp. 237–257). Berlin: Springer. Taghiabadi, E.R., Gromov, V., Fahland, D., & van der Aalst, W.M.P. (2014). Compliance checking of data-aware and resource-aware compliance requirements. In On the move to meaningful internet systems, LNCS 8841 (pp. 237–257). Berlin: Springer.
go back to reference Tax, N., Sidorova, N., van der Aalst, W.M.P., & Haakma, R. (2016). Heuristic approaches for generating local process models through log projections. In Proceedings of IEEE symposium series on computational intelligence (pp. 1–8). Piscataway: IEEE. Tax, N., Sidorova, N., van der Aalst, W.M.P., & Haakma, R. (2016). Heuristic approaches for generating local process models through log projections. In Proceedings of IEEE symposium series on computational intelligence (pp. 1–8). Piscataway: IEEE.
go back to reference Thomas, L.T., Valluri, S.R., & Karlapalem, K. (2010). Margin: maximal frequent subgraph mining. ACM Transactions on Knowledge Discovery from Data, 4(3), 10.CrossRef Thomas, L.T., Valluri, S.R., & Karlapalem, K. (2010). Margin: maximal frequent subgraph mining. ACM Transactions on Knowledge Discovery from Data, 4(3), 10.CrossRef
go back to reference van Beest, N., Dumas, M., García-Bañuelos, L., & La Rosa, M. (2015). Log delta analysis: Interpretable differencing of business process event logs. In Business process management, LNCS 9253 (pp. 386–405). Berlin: Springer. van Beest, N., Dumas, M., García-Bañuelos, L., & La Rosa, M. (2015). Log delta analysis: Interpretable differencing of business process event logs. In Business process management, LNCS 9253 (pp. 386–405). Berlin: Springer.
go back to reference van den Broucke, S.K., Munoz-Gama, J., Carmona, J., Baesens, B., & Vanthienen, J. (2014). Event-based real-time decomposed conformance analysis. In On the move to meaningful internet systems (pp. 345–363). Berlin: Springer. van den Broucke, S.K., Munoz-Gama, J., Carmona, J., Baesens, B., & Vanthienen, J. (2014). Event-based real-time decomposed conformance analysis. In On the move to meaningful internet systems (pp. 345–363). Berlin: Springer.
go back to reference van der Aalst, W.M.P. (2013). Decomposing petri nets for process mining: a generic approach. Distributed and Parallel Databases, 31(4), 471–507.CrossRef van der Aalst, W.M.P. (2013). Decomposing petri nets for process mining: a generic approach. Distributed and Parallel Databases, 31(4), 471–507.CrossRef
go back to reference van der Aalst, W.M.P. (2016). Process mining: data science in action. Berlin: Springer.CrossRef van der Aalst, W.M.P. (2016). Process mining: data science in action. Berlin: Springer.CrossRef
go back to reference van der Aalst, W.M.P., van Dongen, B.F., Herbst, J., Maruster, L., Schimm, G., & Weijters, A.J. (2003). Workflow mining: a survey of issues and approaches. Data & Knowledge Engineering, 47(2), 237–267.CrossRef van der Aalst, W.M.P., van Dongen, B.F., Herbst, J., Maruster, L., Schimm, G., & Weijters, A.J. (2003). Workflow mining: a survey of issues and approaches. Data & Knowledge Engineering, 47(2), 237–267.CrossRef
go back to reference van der Aalst, W.M.P., Adriansyah, A., & van Dongen, B. (2012). Replaying history on process models for conformance checking and performance analysis. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2(2), 182–192. van der Aalst, W.M.P., Adriansyah, A., & van Dongen, B. (2012). Replaying history on process models for conformance checking and performance analysis. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2(2), 182–192.
Metadata
Title
Discovering anomalous frequent patterns from partially ordered event logs
Authors
Laura Genga
Mahdi Alizadeh
Domenico Potena
Claudia Diamantini
Nicola Zannone
Publication date
24-02-2018
Publisher
Springer US
Published in
Journal of Intelligent Information Systems / Issue 2/2018
Print ISSN: 0925-9902
Electronic ISSN: 1573-7675
DOI
https://doi.org/10.1007/s10844-018-0501-z

Other articles of this Issue 2/2018

Journal of Intelligent Information Systems 2/2018 Go to the issue

Premium Partner