Skip to main content
Top

2013 | Book

Advances in Intelligent Data Analysis XII

12th International Symposium, IDA 2013, London, UK, October 17-19, 2013. Proceedings

Editors: Allan Tucker, Frank Höppner, Arno Siebes, Stephen Swift

Publisher: Springer Berlin Heidelberg

Book Series : Lecture Notes in Computer Science

insite
SEARCH

About this book

This book constitutes the refereed conference proceedings of the 12th International Conference on Intelligent Data Analysis, which was held in October 2013 in London, UK. The 36 revised full papers together with 3 invited papers were carefully reviewed and selected from 84 submissions handling all kinds of modeling and analysis methods, irrespective of discipline. The papers cover all aspects of intelligent data analysis, including papers on intelligent support for modeling and analyzing data from complex, dynamical systems.

Table of Contents

Frontmatter

Invited Papers

Data, Not Dogma: Big Data, Open Data, and the Opportunities Ahead

Big data and open data promise tremendous advances. But the media hype ignores the difficulties and the risks associated with this promise. Beginning with the observation that people want

answers to questions

, not simply data, I explore some of the difficulties and risks which lie in the path of realising the opportunities.

David J. Hand
Computational Techniques for Crop Disease Monitoring in the Developing World

Tracking the spread of viral crop diseases is critically important in developing countries. It is also a problem in which several data analysis techniques can be applied in order to get more reliable information more quickly and at lower cost. This paper describes some novel ways in which computer vision, spatial modelling, active learning and optimisation can be applied in this setting, based on experiences of surveying viral diseases affecting cassava and banana crops in Uganda.

John Quinn
Subjective Interestingness in Exploratory Data Mining

Exploratory data mining has as its aim to assist a user in improving their understanding about the data. Considering this aim, it seems self-evident that in optimizing this process the data as well as the user need to be considered. Yet, the vast majority of exploratory data mining methods (including most methods for clustering, itemset and association rule mining, subgroup discovery, dimensionality reduction, etc) formalize interestingness of patterns in an objective manner, disregarding the user altogether. More often than not this leads to subjectively uninteresting patterns being reported.

Here I will discuss a general mathematical framework for formalizing interestingness in a subjective manner. I will further demonstrate how it can be successfully instantiated for a variety of exploratory data mining problems. Finally, I will highlight some connections to other work, and outline some of the challenges and research opportunities ahead.

Tijl De Bie

Selected Contributions

Time Point Estimation of a Single Sample from High Throughput Experiments Based on Time-Resolved Data and Robust Correlation Measures

Recent advances of modern high-throughput technologies such as mass spectrometry and microarrays allow the measurement of cell products like proteins, peptides and mRNA under different conditions over time. Therefore, researchers have to deal with a vast amount of available measurements gained from accomplished experiments using the above techniques.

In this paper, we set our focus on methods that analyze consistency of time-resolved replicates by using similarity patterns between measured cell products over time. This fact led us to develop and evaluate a method for time points estimation of a single sample using independent replicate sets taking the existing noise in the measurements and biological perturbations into account. Moreover, the established approach can be applied to assess the preanalytical quality of biobank samples used in further biomarker research.

Nada Abidi, Frank Klawonn, Jörg Oliver Thumfart
Detecting Events in Molecular Dynamics Simulations

We describe the application of a recently published general event detection framework, called

EVE

to the challenging task of molecular event detection, that is, the automatic detection of structural changes of a molecule over time. Different types of molecular events can be of interest which have, in the past, been addressed by specialized methods. The framework used here allows different types of molecular events to be systematically investigated. In this paper, we summarize existing molecular event detection methods and demonstrate how

EVE

can be configured for a number of molecular event types.

Iris Adä, Michael R. Berthold
Graph Clustering by Maximizing Statistical Association Measures

We are interested in objective functions for clustering undirected and unweighted graphs. Our goal is to define alternatives to the popular modularity measure. To this end, we propose to adapt statistical association coefficients, which traditionally measure the proximity between partitions, for graph clustering. Our approach relies on the representation of statistical association measures in a relational formulation which uses the adjacency matrices of the equivalence relations underlying the partitions. We show that graph clustering can then be solved by fitting the graph with an equivalence relation

via

the maximization of a statistical association coefficient. We underline the connections between the proposed framework and the modularity model. Our theoretical work comes with an empirical study on computer-generated graphs. Our results show that the proposed methods can recover the community structure of a graph similarly or better than the modularity.

Julien Ah-Pine
Evaluation of Association Rule Quality Measures through Feature Extraction

The practical success of association rule mining depends heavily on the criterion to choose among the many rules often mined. Many rule quality measures exist in the literature. We propose a protocol to evaluate the evaluation measures themselves. For each association rule, we measure the improvement in accuracy that a commonly used predictor can obtain from an additional feature, constructed according to the exceptions to the rule. We select a reference set of rules that are helpful in this sense. Then, our evaluation method takes into account both how many of these helpful rules are found near the top rules for a given quality measure, and how near the top they are. We focus on seven association rule quality measures. Our experiments indicate that multiplicative improvement and (to a lesser extent) support and leverage (a.k.a. weighted relative accuracy) tend to obtain better results than the other measures.

José L. Balcázar, Francis Dogbey
Towards Comprehensive Concept Description Based on Association Rules

The paper presents two approaches to post-processing of association rules that are used for concept description. The first approach is based on the idea of meta-learning; a subsequent association rule mining step is applied to the results of ”standard” association rule mining. We thus obtain ”rules about rules” that in a condensed form represent the knowledge found using association rules generated in the first step. The second approach finds a ”core” part of the association rules that can be used to derive the confidence of every rule created in the first step. Again, the core part is substantially smaller than the set of all association rules. We experimentally evaluate the proposed methods on some benchmark data taken from the UCI repository. The system LISp-Miner has been used to carry out the experiments.

Petr Berka
CD-MOA: Change Detection Framework for Massive Online Analysis

Analysis of data from networked digital information systems such as mobile devices, remote sensors, and streaming applications, needs to deal with two challenges: the size of data and the capacity to be adaptive to changes in concept in real-time. Many approaches meet the challenge by using an explicit change detector alongside a classification algorithm and then evaluate performance using classification accuracy. However, there is an unexpected connection between change detectors and classification methods that needs to be acknowledged. The phenomenon has been observed previously, connecting high classification performance with high false positive rates. The implication is that we need to be careful to evaluate systems against intended outcomes–high classification rates, low false alarm rates, compromises between the two and so forth. This paper proposes a new experimental framework for evaluating change detection methods against intended outcomes. The framework is general in the sense that it can be used with other data mining tasks such as frequent item and pattern mining, clustering etc. Included in the framework is a new measure of performance of a change detector that monitors the compromise between fast detection and false alarms. Using this new experimental framework we conduct an evaluation study on synthetic and real-world datasets to show that classification performance is indeed a poor proxy for change detection performance and provide further evidence that classification performance is correlated strongly with the use of change detectors that produce high false positive rates.

Albert Bifet, Jesse Read, Bernhard Pfahringer, Geoff Holmes, Indrė Žliobaitė
Integrating Multiple Studies of Wheat Microarray Data to Identify Treatment-Specific Regulatory Networks

Microarrays have allowed biologists to better understand gene regulatory mechanisms. Wheat microarray data analysis is a complex and challenging topic and knowledge of gene regulation in wheat is still very superficial. However, understanding key mechanisms in this plant holds much potential for food security, especially with a changing climate. The purpose of this paper is to combine multiple microarray studies to automatically identify subnetworks that are distinctive to specific experimental conditions. For example, identifying a regulatory network of genes that only exists under certain types of experimental conditions will assist in understanding the nature of the mechanisms. We derive unique networks from multiple independent networks to better understand key mechanisms and how they change under different conditions. We compare the results with biclustering, detect the most predictive genes and validate the results based upon known biological mechanisms. We also explore how this pipeline performs on yeast microarray data.

Valeria Bo, Artem Lysenko, Mansoor Saqi, Dimah Habash, Allan Tucker
Finding Frequent Patterns in Parallel Point Processes

We consider the task of finding frequent patterns in parallel point processes—also known as finding frequent parallel episodes in event sequences. This task can be seen as a generalization of frequent item set mining: the co-occurrence of items (or events) in transactions is replaced by their (imprecise) co-occurrence on a continuous (time) scale, meaning that they occur in a limited (time) span from each other. We define the support of an item set in this setting based on a maximum independent set approach allowing for efficient computation. Furthermore, we show how the enumeration and test of candidate sets can be made efficient by properly reducing the event sequences and exploiting perfect extension pruning. Finally, we study how the resulting frequent item sets/patterns can be filtered for closed and maximal sets.

Christian Borgelt, David Picado-Muiño
Behavioral Clustering for Point Processes

Groups of (parallel) point processes may be analyzed with a variety of different goals. Here we consider the case in which one has a special interest in finding subgroups of processes showing a behavior that differs significantly from the other processes. In particular, we are interested in finding subgroups that exhibit an increased synchrony. Finding such groups of processes poses a difficult problem as its naïve solution requires enumerating the power set of all processes involved, which is a costly procedure. In this paper we propose a method that allows us to efficiently filter the process set for candidate subgroups. We pay special attention to the possibilities of

temporal imprecision

, meaning that the synchrony is not exact, and

selective participation

, meaning that only a subset of the related processes participates in each synchronous event.

Christian Braune, Christian Borgelt, Rudolf Kruse
Estimating Prediction Certainty in Decision Trees

Decision trees estimate prediction certainty using the class distribution in the leaf responsible for the prediction. We introduce an alternative method that yields better estimates. For each instance to be predicted, our method inserts the instance to be classified in the training set with one of the possible labels for the target attribute; this procedure is repeated for each one of the labels. Then, by comparing the outcome of the different trees, the method can identify instances that might present some difficulties to be correctly classified, and attribute some uncertainty to their prediction. We perform an extensive evaluation of the proposed method, and show that it is particularly suitable for ranking and reliability estimations. The ideas investigated in this paper may also be applied to other machine learning techniques, as well as combined with other methods for prediction certainty estimation.

Eduardo P. Costa, Sicco Verwer, Hendrik Blockeel
Interactive Discovery of Interesting Subgroup Sets

Although subgroup discovery aims to be a practical tool for exploratory data mining, its wider adoption is hampered by redundancy and the re-discovery of common knowledge. This can be remedied by parameter tuning and manual result filtering, but this requires considerable effort from the data analyst. In this paper we argue that it is essential to involve the user in the discovery process to solve these issues. To this end, we propose an interactive algorithm that allows a user to provide feedback

during search

, so that it is steered towards more interesting subgroups. Specifically, the algorithm exploits user feedback to guide a diverse beam search. The empirical evaluation and a case study demonstrate that uninteresting subgroups can be effectively eliminated from the results, and that the overall effort required to obtain interesting and diverse subgroup sets is reduced. This confirms that within-search interactivity can be useful for data analysis.

Vladimir Dzyuba, Matthijs van Leeuwen
Gaussian Mixture Models for Time Series Modelling, Forecasting, and Interpolation

Gaussian mixture models provide an appealing tool for time series modelling. By embedding the time series to a higher-dimensional space, the density of the points can be estimated by a mixture model. The model can directly be used for short-to-medium term forecasting and missing value imputation. The modelling setup introduces some restrictions on the mixture model, which when appropriately taken into account result in a more accurate model. Experiments on time series forecasting show that including the constraints in the training phase particularly reduces the risk of overfitting in challenging situations with missing values or a large number of Gaussian components.

Emil Eirola, Amaury Lendasse
When Does Active Learning Work?

Active Learning (AL) methods seek to improve classifier performance when labels are expensive or scarce. We consider two central questions: Where does AL work? How much does it help? To address these questions, a comprehensive experimental simulation study of Active Learning is presented. We consider a variety of tasks, classifiers and other AL factors, to present a broad exploration of AL performance in various settings. A precise way to quantify performance is needed in order to know when AL works. Thus we also present a detailed methodology for tackling the complexities of assessing AL performance in the context of this experimental study.

Lewis P. G. Evans, Niall M. Adams, Christoforos Anagnostopoulos
OrderSpan: Mining Closed Partially Ordered Patterns

Due to the complexity of the task, partially ordered pattern mining of sequential data has not been subject to much study, despite its usefulness. This paper investigates this data mining challenge by describing OrderSpan, a new algorithm that extracts such patterns from sequential databases and overcomes some of the drawbacks of existing methods. Our work consists in providing a simple and flexible framework to directly mine complex sequences of itemsets, by combining well-known properties on prefixes and suffixes. Experiments were performed on different real datasets to show the benefit of partially ordered patterns.

Mickaël Fabrègue, Agnès Braud, Sandra Bringay, Florence Le Ber, Maguelonne Teisseire
Learning Multiple Temporal Matching for Time Series Classification

In real applications, time series are generally of complex structure, exhibiting different global behaviors within classes. To discriminate such challenging time series, we propose a multiple temporal matching approach that reveals the commonly shared features within classes, and the most differential ones across classes. For this, we rely on a new framework based on the variance/covariance criterion to strengthen or weaken matched observations according to the induced variability within and between classes. The experiments performed on real and synthetic datasets demonstrate the ability of the multiple temporal matching approach to capture fine-grained distinctions between time series.

Cedric Frambourg, Ahlame Douzal-Chouakria, Eric Gaussier
On the Importance of Nonlinear Modeling in Computer Performance Prediction

Computers are nonlinear dynamical systems that exhibit complex and sometimes even chaotic behavior. The low-level performance models used in the computer systems community, however, are linear. This paper is an exploration of that disconnect: when linear models are adequate for predicting computer performance and when they are not. Specifically, we build linear and nonlinear models of the processor load of an Intel i7-based computer as it executes a range of different programs. We then use those models to predict the processor loads forward in time and compare those forecasts to the true continuations of the time series.

Joshua Garland, Elizabeth Bradley
Diversity-Driven Widening

This paper follows our earlier publication [1], where we introduced the idea of tuned data mining which draws on parallel resources to improve model accuracy rather than the usual focus on speed-up. In this paper we present a more in-depth analysis of the concept of Widened Data Mining, which aims at reducing the impact of greedy heuristics by exploring more than just one suitable solution at each step. In particular we focus on how diversity considerations can substantially improve results. We again use the greedy algorithm for the set cover problem to demonstrate these effects in practice.

Violeta N. Ivanova, Michael R. Berthold
Towards Indexing of Web3D Signing Avatars

Signing avatars are becoming common and being published on the World Wide Web at an incredible rate. They are actually used in education to help deaf children and their parents to learn sign language thanks to many 3D signing avatars systems. In Tunisia, signing avatars have been used since several years as part of the WebSign project. During the last few years, thousands of 3D signing avatars have been recorded using WebSign and few other systems. One of the major challenges that we was facing is how to index and retrieve efficiently this huge quantity of 3D signed scenes. Indexing and cataloging these signed scenes is beyond the capabilities of current text-based search engines. In this paper, we describe a system that collects 3D signed scenes, processes and recognizes the signed words inside them. The processed scenes are then indexed for later retrieval. We use a novel approach for sign language recognition from 3D scenes based on the Longest Common Subsequence algorithm. Our system is able to recognize signs inside 3D scenes at a rate of 96.5 % using a 3D model dictionary. We present also a novel approach for search and results ranking based on the similarity rates between 3D models. Our method is more efficient than Hidden Markov Models in term of recognition time and in the case of co-articulated signs.

Kabil Jaballah, Mohamed Jemni
Variational Bayesian PCA versus k-NN on a Very Sparse Reddit Voting Dataset

We present vote estimation results on the largely unexplored Reddit voting dataset that contains 23M votes from 43k users on 3.4M links. This problem is approached using Variational Bayesian Principal Component Analysis (VBPCA) and a novel algorithm for

k

-Nearest Neighbors (

k

-NN) optimized for high dimensional sparse datasets without using any approximations. We also explore the scalability of the algorithms for extremely sparse problems with up to 99.99% missing values. Our experiments show that

k

-NN works well for the standard Reddit vote prediction. The performance of VBPCA with the full preprocessed Reddit data was not as good as

k

-NN’s, but it was more resilient to further sparsification of the problem.

Jussa Klapuri, Ilari Nieminen, Tapani Raiko, Krista Lagus
Analysis of Cluster Structure in Large-Scale English Wikipedia Category Networks

In this paper we propose a framework for analysing the structure of a large-scale social media network, a topic of significant recent interest. Our study is focused on the Wikipedia category network, where nodes correspond to Wikipedia categories and edges connect two nodes if the nodes share at least one common page within the Wikipedia network. Moreover, each edge is given a weight that corresponds to the number of pages shared between the two categories that it connects. We study the structure of category clusters within the three complete English Wikipedia category networks from 2010 to 2012. We observe that category clusters appear in the form of well-connected components that are naturally clustered together. For each dataset we obtain a graph, which we call the

t-filtered

category graph, by retaining just a single edge linking each pair of categories for which the weight of the edge exceeds some specified threshold

t

. Our framework exploits this graph structure and identifies connected components within the

t-filtered

category graph. We studied the large-scale structural properties of the three Wikipedia category networks using the proposed approach. We found that the number of categories, the number of clusters of size two, and the size of the largest cluster within the graph all appear to follow power laws in the threshold

t

. Furthermore, for each network we found the value of the threshold

t

for which increasing the threshold to

t

 + 1 caused the “giant” largest cluster to diffuse into two or more smaller clusters of significant size and studied the semantics behind this diffusion.

Thidawan Klaysri, Trevor Fenner, Oded Lachish, Mark Levene, Panagiotis Papapetrou
1d-SAX: A Novel Symbolic Representation for Time Series

SAX (Symbolic Aggregate approXimation) is one of the main symbolization techniques for time series. A well-known limitation of SAX is that trends are not taken into account in the symbolization. This paper proposes 1d-SAX a method to represent a time series as a sequence of symbols that each contain information about the average and the trend of the series on a segment. We compare the efficiency of SAX and 1d-SAX in terms of goodness-of-fit, retrieval and classification performance for querying a time series database with an asymmetric scheme. The results show that 1d-SAX improves performance using equal quantity of information, especially when the compression rate increases.

Simon Malinowski, Thomas Guyet, René Quiniou, Romain Tavenard
Learning Models of Activities Involving Interacting Objects

We propose the LEMAIO multi-layer framework, which makes use of hierarchical abstraction to learn models for activities involving multiple interacting objects from time sequences of data concerning the individual objects. Experiments in the sea navigation domain yielded learned models that were then successfully applied to activity recognition, activity simulation and multi-target tracking. Our method compares favourably with respect to previously reported results using Hidden Markov Models and Relational Particle Filtering.

Cristina Manfredotti, Kim Steenstrup Pedersen, Howard J. Hamilton, Sandra Zilles
Correcting the Usage of the Hoeffding Inequality in Stream Mining

Many stream classification algorithms use the Hoeffding Inequality to identify the best split attribute during tree induction.

We show that the prerequisites of the Inequality are violated by these algorithms, and we propose corrective steps. The new stream classification core,

correctedVFDT

, satisfies the prerequisites of the Hoeffding Inequality and thus provides the expected performance guarantees.

The goal of our work is not to improve accuracy, but to guarantee a reliable and interpretable error bound. Nonetheless, we show that our solution achieves lower error rates regarding split attributes and sooner split decisions while maintaining a similar level of accuracy.

Pawel Matuszyk, Georg Krempl, Myra Spiliopoulou
Exploratory Data Analysis through the Inspection of the Probability Density Function of the Number of Neighbors

Exploratory data analysis is a fundamental stage in data mining of high-dimensional datasets. Several algorithms have been implemented to grasp a general idea of the geometry and patterns present in high-dimensional data. Here, we present a methodology based on the distance matrix of the input data. The algorithm is based in the number of points considered to be neighbors of each input vector. Neighborhood is defined in terms of an hypersphere of varying radius, and from the distance matrix the probability density function of the number of neighbor vectors is computed. We show that when the radius of the hypersphere is systematically increased, a detailed analysis of the probability density function of the number of neighbors unfolds relevant aspects of the overall features that describe the high-dimensional data. The algorithm is tested with several datasets and we show its pertinence as an exploratory data analysis tool.

Antonio Neme, Antonio Nido
The Modelling of Glaucoma Progression through the Use of Cellular Automata

We propose a model of glaucoma progression based on the application of Cellular Automata (CA) to visual field (VF) data, obtained through automated perimetry. VF sensitivities are converted into ganglion cell loss and CA are utilised to model the gradual deterioration of vision, mimicking degeneration of the actual ganglia. First we discuss the construction of a grid that approximates the VF map and the corresponding layer of ganglia in terms of cell counts in individual fields. The grid is populated with dead cells in accordance with patients’ tests, and then we run a CA, utilising a majority and a probabilistic rule. Preliminary results are presented, showing that during its evolution, the CA often converges to configurations where the death of cells resembles VF data of the same patients, at later time. That is, the percentage loss of cells in VF fields observed in the CA resembles the real VF data.

Stelios Pavlidis, Stephen Swift, Allan Tucker, Steve Counsell
Towards Narrative Ideation via Cross-Context Link Discovery Using Banded Matrices

Knowledge discovery and computational creativity have until lately been investigated by two separate research communities. However, research in bisociative, cross-context knowledge discovery has recently started addressing creative tasks, including creative literature mining. This paper contributes to this effort by investigating an approach to cross-context link discovery based on banded matrices, aimed at identifying meaningful bridging terms (b-terms) at the intersection of two different domains. The proposed approach was applied to a simplified computational creativity task of narrative ideation from pairs of short sentences. As input, we took sentences from two different contexts: what-if sentences retrieved from Twitter, and morals from Aesop’s fables, respectively. The approach resulted in a list of linked pairs of sentences from these two domains, illustrating the potential of the proposed approach to cross-context narrative ideation.

Matic Perovšek, Bojan Cestnik, Tanja Urbančič, Simon Colton, Nada Lavrač
Gaussian Topographic Co-clustering Model

The visualization of the clusters obtained by a partitioning procedure is very informative as this helps to a better overview of the contents of a data table. For co-clustering, the latent block mixture model is very effective. We propose to define generative self-organizing maps with this model for Gaussian blocks. A perspective is the analysis and the visualization of continuous data.

Rodolphe Priam, Mohamed Nadif, Gérard Govaert
Preventing Churn in Telecommunications: The Forgotten Network

This paper outlines an approach developed as a part of a company-wide churn management initiative of a major European telecom operator. We are focusing on explanatory churn model for the postpaid segment, assuming that the mobile telecom network, the key resource of operators, is also a churn driver in case it under delivers to customers’ expectations. Typically, insights generated by churn models are deployed in marketing campaigns; our model’s insights are used in network optimization in order to remove the key network related churn drivers and therefore prevent churn, rather than cure it. The insights generated by the model have caused a paradigm shift in managing the network with the operator where the research was conducted.

Dejan Radosavljevik, Peter van der Putten
Computational Properties of Fiction Writing and Collaborative Work

From the earliest days of computing, there have been tools to help shape narrative. Spell-checking, word counts, and readability analysis, give today’s novelists tools that Dickens, Austen, and Shakespeare could only have dreamt of. However, such tools have focused on the word, or phrase levels. In the last decade, research focus has shifted to support for collaborative editing of documents. This work considers more sophisticated attempts to visualise the semantics, pace and rhythm within a narrative through data mining. We describe real life applications in two related domains.

Joseph Reddington, Fionn Murtagh, Douglas Cowie
Classifier Evaluation with Missing Negative Class Labels

The concept of a negative class does not apply to many problems for which classification is increasingly utilized. In this study we investigate the reliability of evaluation metrics when the negative class contains an unknown proportion of mislabeled positive class instances. We examine how evaluation metrics can inform us about potential systematic biases in the data. We provide a motivating case study and a general framework for approaching evaluation when the negative class contains mislabeled positive class instances. We show that the behavior of evaluation metrics is unstable in the presence of uncertainty in class labels and that the stability of evaluation metrics depends on the kind of bias in the data. Finally, we show that the type and amount of bias present in data can have a significant effect on the ranking of evaluation metrics and the degree to which they over- or underestimate the true performance of classifiers.

Andrew K. Rider, Reid A. Johnson, Darcy A. Davis, T. Ryan Hoens, Nitesh V. Chawla
Dynamic MMHC: A Local Search Algorithm for Dynamic Bayesian Network Structure Learning

Dynamic Bayesian networks (DBNs) are a class of probabilistic graphical models that has become a standard tool for modeling various stochastic time-varying phenomena. Probabilistic graphical models such as 2-Time slice BN (2T-BNs) are the most used and popular models for DBNs. Because of the complexity induced by adding the temporal dimension, DBN structure learning is a very complex task. Existing algorithms are adaptations of score-based BN structure learning algorithms but are often limited when the number of variables is high. We focus in this paper to DBN structure learning with another family of structure learning algorithms, local search methods, known for its scalability. We propose Dynamic MMHC, an adaptation of the ”static” MMHC algorithm. We illustrate the interest of this method with some experimental results.

Ghada Trabelsi, Philippe Leray, Mounir Ben Ayed, Adel Mohamed Alimi
Accurate Visual Features for Automatic Tag Correction in Videos

We present a new system for video auto tagging which aims at correcting the tags provided by users for videos uploaded on the Internet. Unlike most existing systems, in our proposal, we do not use the questionable textual information nor any supervised learning system to perform a tag propagation. We propose to compare directly the visual content of the videos described by different sets of features such as Bag-Of-visual-Words or frequent patterns built from them. We then propose an original tag correction strategy based on the frequency of the tags in the visual neighborhood of the videos. Experiments on a Youtube corpus show that our method can effectively improve the existing tags and that frequent patterns are useful to construct accurate visual features.

Hoang-Tung Tran, Elisa Fromont, François Jacquenet, Baptiste Jeudy
Ontology Database System and Triggers

An ontology database system is a basic relational database management system that models an ontology plus its instances. To reason over the transitive closure of instances in the subsumption hierarchy, an ontology database can either unfold views at query time or propagate assertions using triggers at load time. In this paper, we present a method to embed ontology knowledge into a relational database through triggers. We demonstrate that by forward computing inferences, we improve query time. We find that: first, ontology database systems scale well for small and medium sized ontologies; and second, ontology database systems are able to answer ontology-based queries deductively; We apply this method to a Glass Identification Ontology, and discuss applications in Neuroscience.

Angelina A. Tzacheva, Tyrone S. Toland, Peyton H. Poole, Daniel J. Barnes
A Policy Iteration Algorithm for Learning from Preference-Based Feedback

Conventional approaches to reinforcement learning assume availability of a numerical feedback signal, but in many domains, this is difficult to define or not available at all. The recently proposed framework of preference-based reinforcement learning relaxes this condition by replacing the quantitative reward signal with qualitative preferences over trajectories. In this paper, we show how to estimate preferences over actions from preferences over trajectories. These action preferences can then be used to learn a preferred policy. The performance of this new approach is evaluated by a comparison with SARSA in three common reinforcement learning benchmark problems, namely mountain car, inverted pendulum, and acrobot. The results are showing convergence rates that are comparable, but achieved with a much less time consuming tuning of the setup.

Christian Wirth, Johannes Fürnkranz
Multiclass Learning from Multiple Uncertain Annotations

Annotating a dataset is one of the major bottlenecks in supervised learning tasks, as it can be expensive and time-consuming. Instead, with the development of crowdsourcing services, it has become easy and fast to collect labels from multiple annotators. Our contribution in this paper is to propose a Bayesian probabilistic approach integrating annotator’s uncertainty in the task of learning from multiple noisy annotators (annotators who generate errors). Furthermore, unlike previous work, our proposed approach is directly formulated to handle categorical labels. This is an important point as real-world datasets often have multiple classes available. Extensive experiments on datasets validate the effectiveness of our approach against previous efficient algorithms.

Chirine Wolley, Mohamed Quafafou
Learning Compositional Hierarchies of a Sensorimotor System

We address the problem of learning static spatial representation of a robot motor system and the environment to solve a general forward/inverse kinematics problem. The latter proves complex for high degree-of-freedom systems. The proposed architecture relates to a recent research in cognitive science, which provides a solid evidence that perception and action share common neural architectures. We propose to model both a motor system and an environment with

compositional hierarchies

and develop an algorithm for learning them together with a mapping between the two. We show that such a representation enables efficient learning and inference of robot states. We present our experiments in a simulated environment and with a humanoid robot Nao.

Jure Žabkar, Aleš Leonardis
Backmatter
Metadata
Title
Advances in Intelligent Data Analysis XII
Editors
Allan Tucker
Frank Höppner
Arno Siebes
Stephen Swift
Copyright Year
2013
Publisher
Springer Berlin Heidelberg
Electronic ISBN
978-3-642-41398-8
Print ISBN
978-3-642-41397-1
DOI
https://doi.org/10.1007/978-3-642-41398-8

Premium Partner