Skip to main content

2014 | Buch

Research and Development in Intelligent Systems XXXI

Incorporating Applications and Innovations in Intelligent Systems XXII

insite
SUCHEN

Über dieses Buch

The papers in this volume are the refereed papers presented at AI-2014, the Thirty-fourth SGAI International Conference on Innovative Techniques and Applications of Artificial Intelligence, held in Cambridge in December 2014 in both the technical and the application streams.

They present new and innovative developments and applications, divided into technical stream sections on Knowledge Discovery and Data Mining, Machine Learning, and Agents, Ontologies and Genetic Programming, followed by application stream sections on Evolutionary Algorithms/Dynamic Modelling, Planning and Optimisation, and Machine Learning and Data Mining. The volume also includes the text of short papers presented as posters at the conference.

This is the thirty-first volume in the Research and Development in Intelligent Systems series, which also incorporates the twenty-second volume in the Applications and Innovations in Intelligent Systems series. These series are essential reading for those who wish to keep up to date with developments in this important field.

Inhaltsverzeichnis

Frontmatter

Research and Development in Intelligent Systems XXXI Best Technical Paper

Frontmatter
On Ontological Expressivity and Modelling Argumentation Schemes Using COGUI
Abstract
Knowledge elicitation, representation and reasoning explanation by/to non computing experts has always been considered as a crafty task due to difficulty of expressing logical statements by non logicians. In this paper, we use the COGUI editor in order to elicit and represent Argumentation Schemes within an inconsistent knowledge base. COGUI is a visual, graph based knowledge representation editor compatible with main Semantic Web languages. COGUI allows for default reasoning on top of ontologies. We investigate its use for modelling and reasoning using Argumentation Schemes and discuss the advantages of such representation. We show how this approach can be useful in the practical setting of EcoBioCap where the different Argumentation Schemes can be used to lead reasoning.
Wael Hamdan, Rady Khazem, Ghaida Rebdawi, Madalina Croitoru, Alain Gutierrez, Patrice Buche

Knowledge Discovery and Data Mining

Frontmatter
Computationally Efficient Rule-Based Classification for Continuous Streaming Data
Abstract
Advances in hardware and software technologies allow to capture streaming data. The area of Data Stream Mining (DSM) is concerned with the analysis of these vast amounts of data as it is generated in real-time. Data stream classification is one of the most important DSM techniques allowing to classify previously unseen data instances. Different to traditional classifiers for static data, data stream classifiers need to adapt to concept changes (concept drift) in the stream in real-time in order to reflect the most recent concept in the data as accurately as possible. A recent addition to the data stream classifier toolbox is eRules which induces and updates a set of expressive rules that can easily be interpreted by humans. However, like most rule-based data stream classifiers, eRules exhibits a poor computational performance when confronted with continuous attributes. In this work, we propose an approach to deal with continuous data effectively and accurately in rule-based classifiers by using the Gaussian distribution as heuristic for building rule terms on continuous attributes. We show on the example of eRules that incorporating our method for continuous attributes indeed speeds up the real-time rule induction process while maintaining a similar level of accuracy compared with the original eRules classifier. We termed this new version of eRules with our approach G-eRules.
Thien Le, Frederic Stahl, João Bártolo Gomes, Mohamed Medhat Gaber, Giuseppe Di Fatta
Improved Stability of Feature Selection by Combining Instance and Feature Weighting
Abstract
The current study presents a technique that aims at improving stability of feature subset selection by means of a combined instance and feature weighting process. Both types of weights are based on margin concepts and can therefore be naturally interlaced. We report experiments performed on both synthetic and real data (including microarray data) showing improvements in selection stability at a similar level of prediction performance.
Gabriel Prat, Lluís A. Belanche
Towards a Parallel Computationally Efficient Approach to Scaling Up Data Stream Classification
Abstract
Advances in hardware technologies allow to capture and process data in real-time and the resulting high throughput data streams require novel data mining approaches. The research area of Data Stream Mining (DSM) is developing data mining algorithms that allow us to analyse these continuous streams of data in real-time. The creation and real-time adaption of classification models from data streams is one of the most challenging DSM tasks. Current classifiers for streaming data address this problem by using incremental learning algorithms. However, even so these algorithms are fast, they are challenged by high velocity data streams, where data instances are incoming at a fast rate. This is problematic if the applications desire that there is no or only a very little delay between changes in the patterns of the stream and absorption of these patterns by the classifier. Problems of scalability to Big Data of traditional data mining algorithms for static (non streaming) datasets have been addressed through the development of parallel classifiers. However, there is very little work on the parallelisation of data stream classification techniques. In this paper we investigate K-Nearest Neighbours (KNN) as the basis for a real-time adaptive and parallel methodology for scalable data stream classification tasks.
Mark Tennant, Frederic Stahl, Giuseppe Di Fatta, João Bártolo Gomes

Machine Learning

Frontmatter
Following the Trail of Source Languages in Literary Translations
Abstract
We build on past research in distinguishing English translations from originally English text, and in guessing the source language where the text is deemed to be a translation. We replicate an extant method in relation to both a reconstruction of the original data set and a fresh data set compiled on an analogous basis. We extend this with an analysis of the features that emerge from the combined data set. Finally, we report on an inverse use of the method, not as guessing the source language of a translated text, but as a tool in quality estimation, marking a text as requiring inspection if it is guessed to be a translation, rather than a text composed originally in the language analysed. We obtain c. 80 % accuracy, comparable to results of earlier work in literary source language guessing—this supports the claim of the method’s validity in identifying salient features of source language interference.
Carmen Klaussner, Gerard Lynch, Carl Vogel
Reluctant Reinforcement Learning
Abstract
This paper presents an approach to Reinforcement Learning that seems to work very well in changing environments. The experiments are based on an unmanned vehicle problem where the vehicle is equipped with navigation cameras and uses a multilayer perceptron (MLP). The route can change and obstacles can be added without warning. In the steady state, no learning takes place, but the system maintains a small cache of recent inputs and rewards. When a negative reward occurs, learning restarts, based not on the immediate situation but on the memory that has generated the greatest error, and the updated strategy is quickly reviewed using the cache of recent memories within an accelerated learning phase. In the resulting Reluctant Learning algorithm the multiple use of a small quantity of previous experiences to validate updates to the strategy moves the MLP towards convergence and finds a balance between exploration of improvements to strategy and exploitation of previous learning.
Chris Jones, Malcolm Crowe
Preference and Sentiment Guided Social Recommendations with Temporal Dynamics
Abstract
Capturing users’ preference that change over time is a great challenge in recommendation systems. What makes a product feature interesting now may become the accepted standard in the future. Social recommender systems that harness knowledge from user expertise and interactions to provide recommendation have great potential in capturing such trending information. In this paper, we model our recommender system using sentiment rich user generated product reviews and temporal information. Specifically we integrate these two resources to formalise a novel aspect-based sentiment ranking that captures temporal distribution of aspect sentiments and so the preferences of the users over time. We demonstrate the utility of our proposed model by conducting a comparative analysis on data extracted from Amazon.com and Cnet. We show that considering the temporal preferences of users leads to better recommendation and that user preferences change over time.
Xavier Ferrer, Yoke Yie Chen, Nirmalie Wiratunga, Enric Plaza

Agents, Ontologies and Genetic Programming

Frontmatter
Query Failure Explanation in Inconsistent Knowledge Bases: A Dialogical Approach
Abstract
In the EcoBioCap project (www.ecobiocap.eu) about the next generation of packaging, a decision support system has been built that uses argumentation to deal with stakeholder preferences. However, when testing the tool the domain experts did not always understand the output of the system. The approach developed in this paper is the first step to the construction of a decision support system endowed with an explanation module. We place ourselves in the equivalent setting of inconsistent Ontology-Based Data Access (OBDA) and addresses the problem of explaining Boolean Conjunctive Query (BCQ) failure. Our proposal relies on an interactive and argumentative approach where the processes of explanation takes the form of a dialogue between the User and the Reasoner. We exploit the equivalence between argumentation and inconsistency tolerant semantics to prove that the Reasoner can always provide an answer for user’s questions.
Abdallah Arioua, Nouredine Tamani, Madalina Croitoru, Patrice Buche
Benchmarking Grammar-Based Genetic Programming Algorithms
Abstract
The publication of Grammatical Evolution (GE) led to the development of numerous variants of this Grammar-Based approach to Genetic Programming (GP). In order for these variants to be compared, the community requires a rigorous means for benchmarking the algorithms. However, GP as a field is not known for the quality of its benchmarking, with many identified problems, making direct comparisons either difficult or impossible in many cases. Aside from there not being a single, agreed-upon, benchmarking test, the tests currently utilised have shown a lack of standardisation. We motivate the following research by identifying some key issues with current benchmarking approaches. We then propose a standardised set of metrics for future benchmarking and demonstrate the use of these metrics by running a comparison of three Grammar-Based Genetic Programming methods. We conclude the work by discussing the results and proposing directions for future benchmarking.
Christopher J. Headleand, Llyr Ap Cenydd, William J. Teahan
The Effects of Bounding Rationality on the Performance and Learning of CHREST Agents in Tileworld
Abstract
Learning in complex and complicated domains is fundamental to performing suitable and timely actions within them. The ability of chess masters to learn and recall huge numbers of board configurations to produce near-optimal actions provides evidence that chunking mechanisms are likely to underpin human learning. Cognitive theories based on chunking argue in favour for the notion of bounded rationality since relatively small chunks of information are learnt in comparison to the total information present in the environment. CHREST, a computational architecture that implements chunking theory, has previously been used to investigate learning in deterministic environments such as chess, where future states are solely dependent upon the actions of agents. In this paper, the CHREST architecture is implemented in agents situated in “Tileworld”, a stochastic environment whose future state depends on both the actions of agents and factors intrinsic to the environment which agents have no control over. The effects of bounding agents’ visual input on learning and performance in various scenarios where the complexity of Tileworld is altered is analysed using computer simulations. Our results show that interactions between independent variables are complex and have important implications for agents situated in stochastic environments where a balance must be struck between learning and performance.
Martyn Lloyd-Kelly, Peter C. R. Lane, Fernand Gobet

Short Papers

Frontmatter
An Autopoietic Repertoire
Abstract
This paper presents a strategy for natural language processing in natural language. Using a concept as a unit of conversation, defined by a repertoire of phrases, it describes the concept of autopoiesis: a repertoire for the construction of repertoires. A minimal repertoire, representing a shopping list app, is described. This is followed by a specification of the autopoietic repertoire, followed by the full repertoire of the shopping list. The advantages of this approach is basically two-fold: a natural language specification is self-evident; moreover, it results in a rich, tiered interface of repertoires supporting repertoires. This paper is validated by an aural information system, publicly available on various mobile platforms.
M. J. Wheatman
De-risking Fleet Replacement Decisions
Abstract
This paper outlines the different modelling approaches for realizing sustainable operations of asset replacement. We study the fleet portfolio management problem that could be faced by a firm deciding ahead which vehicles to choose for its fleet. In particular it suggests a model that enables generating a plan of vehicle replacement actions with cost minimization and risk exposure simulation. It proposes to use conditional value at risk (CVaR) to account for uncertainty in the decision process, and to use clusters modelling to align the generated plan with vehicle utilization.
Anne Liret, Amir H. Ansaripoor, Fernando S. Oliveira
Reliability and Effectiveness of Cross-Validation in Feature Selection
Abstract
Feature selection is increasingly important in data analysis and machine learning in the big data era. However, how to use the data in feature selection has become a serious issue as the conventional practice of using ALL the data in feature selection may lead to selection bias and some suggest to use PART of the data instead. This paper investigates the reliability and effectiveness of a PART approach implemented by cross validation mechanism in feature selection filters and compares it with the ALL approach. The reliability is measured by an Inter-system Average Tanimoto Index and the effectiveness of the selected features is measured by the mean generalisation accuracy of classification. The experiments are carried out by using synthetic datasets generated with a fixed number of relevant features and varied numbers of irrelevant features and instances, and different level of noise, to mimic some possible real world environments. The results indicate that the PART approach is more effective in reducing the bias when the dataset is small but starts to lose its advantage as the dataset size increases.
Ghadah Aldehim, Wenjia Wang
Self-reinforced Meta Learning for Belief Generation
Abstract
Contrary to common perception, learning does not stop once knowledge has been transferred to an agent. Intelligent behaviour observed in humans and animals strongly suggests that after learning, we self-organise our experiences and knowledge, so that they can be more efficiently reused; a process that is unsupervised and employs reasoning based on the acquired knowledge. Our proposed algorithm emulates meta-learning in-silico: creating beliefs from previously acquired knowledge representations, which in turn become subject to learning, and are further self-reinforced. The proposition of meta-learning, in the form of an algorithm that can learn how to create beliefs on its own accord, raises an interesting question: can artificial intelligence arrive to similar beliefs, rules or ideas, as the ones we humans come to? The described work briefly analyses existing theories and research, and formalises a practical implementation of a meta-learning algorithm.
Alexandros Gkiokas, Alexandra I. Cristea, Matthew Thorpe

Applications and Innovations in Intelligent Systems XXII Best Application Paper

Frontmatter
Combining Semantic Web Technologies with Evolving Fuzzy Classifier eClass for EHR-Based Phenotyping: A Feasibility Study
Abstract
In parallel to nation-wide efforts for setting up shared electronic health records (EHRs) across healthcare settings, several large-scale national and international projects are developing, validating, and deploying electronic EHR-oriented phenotype algorithms that aim at large-scale use of EHRs data for genomic studies. A current bottleneck in using EHRs data for obtaining computable phenotypes is to transform the raw EHR data into clinically relevant features. The research study presented here proposes a novel combination of Semantic Web technologies with the on-line evolving fuzzy classifier eClass to obtain and validate EHR-driven computable phenotypes derived from 1,956 clinical statements from EHRs. The evaluation performed with clinicians demonstrates the feasibility and practical acceptability of the approach proposed.
M. Arguello, S. Lekkas, J. Des, M.J. Fernandez-Prieto, L. Mikhailov

Evolutionary Algorithms/Dynamic Modelling

Frontmatter
Rail-Freight Crew Scheduling with a Genetic Algorithm
Abstract
This article presents a novel genetic algorithm designed for the solution of the Crew Scheduling Problem (CSP) in the rail-freight industry. CSP is the task of assigning drivers to a sequence of train trips while ensuring that no driver’s schedule exceeds the permitted working hours, that each driver starts and finishes their day’s work at the same location, and that no train routes are left without a driver. Real-life CSPs are extremely complex due to the large number of trips, opportunities to use other means of transportation, and numerous government regulations and trade union agreements. CSP is usually modelled as a set-covering problem and solved with linear programming methods. However, the sheer volume of data makes the application of conventional techniques computationally expensive, while existing genetic algorithms often struggle to handle the large number of constraints. A genetic algorithm is presented that overcomes these challenges by using an indirect chromosome representation and decoding procedure. Experiments using real schedules on the UK national rail network show that the algorithm provides an effective solution within a faster timeframe than alternative approaches.
E. Khmeleva, A. A. Hopgood, L. Tipi, M. Shahidan
CR-Modified SOM to the Problem of Handwritten Digits Recognition
Abstract
Recently, researchers show that the handwritten digit recognition is a challenging problem. In this paper first, we introduce a Modified Self Organizing Maps for vector quantization problem then we present a Convolutional Recursive Modified SOM to the problem of handwritten digit recognition. The Modified SOM is novel in the sense of initialization process and the topology preservation. The experimental result on the well known digit database of MNIST, denotes the superiority of the proposed algorithm over the existing SOM-based methods.
Ehsan Mohebi, Adil Bagirov
Dynamic Place Profiles from Geo-folksonomies on the GeoSocial Web
Abstract
The growth of the Web and the increase in using GPS-enabled devices, coupled with the exponential growth of the social media sites, have led to a surge in research interest in Geo-folksonomy analysis. In Geo-Folksonomy, a user assigns an electronic tag to a geographical place resource identified by its longitude and latitude. The assigned tags are used to manage, categorize and describe place resources. Building data models of Geo-folksonomy data sets that represents and analyses, tags, location and time information can be helpful in studying and analysing place information. The aim of my research is to use the spatio temporal data available on the Social web, to extract dynamic place profile. Building a dynamic profile involves including the temporal dimension in the Geo-folksonomy. Indeed, adding the temporal dimension can provide an understanding of geographic places as perceived by users over time.
Soha Mohamed, Alia Abdelmoty

Planning and Optimisation

Frontmatter
Hierarchical Type-2 Fuzzy Logic Based Real Time Dynamic Operational Planning System
Abstract
Operational resource planning is critical for successful operations in service-based organizations as it underpins the process of utilizing resources to achieve a higher quality of service whilst lowering operational costs. The majority of service-based organizations use static operational planning. In recent years these, organizations have made attempts to switch to dynamic operational planners with the view of generating real-time operational plans. This paper proposes a hierarchical type-2 fuzzy logic based operational planner that can work in dynamic environments and can maintain operational plans in real-time. The proposed system outperformed ordinary heuristic-based systems and task dispatchers.
Ahmed Mohamed, Hani Hagras, Sid Shakya, Anne Liret, Raphael Dorne, Gilbert Owusu
A Hybrid Algorithm for Solving Gate Assignment Problem with Robustness and Tow Considerations
Abstract
In this paper, we propose a new method to schedule and evaluate the airport gate assignment. A mathematical model is built to show the relationship among three factors impacting on the gate assignment: robustness, tows, and passenger transfer distance. The stochastic uncertainty parameter is estimated by analyzing the historical data in Hong Kong International Airport (HKIA). Finally, an Artificial Intelligence-based hybrid meta-heuristic is designed to solve this problem.
C. H. Yu, Henry Y. K. Lau
An Iterative Heuristics Algorithm for Solving the Integrated Aircraft and Passenger Recovery Problem
Abstract
Airline disruption incurred huge cost for airlines and serious inconvenience for travelers. In this paper, we study the integrated aircraft and passenger schedule recovery problem. To efficiently solve this problem, we proposed decomposition method to divide the whole problem into two smaller problems. An iterative heuristics strategy is proposed to improve solution quality by iteratively solving decomposed problems. Our algorithm is tested on the data set provided by ROADEF 2009. We simulate several airport closure scenarios and experimental results show that our algorithm can provide a high quality solution in the required time limit.
Zhang Dong, H.Y.K. Henry Lau

Machine Learning and Data Mining

Frontmatter
A Framework for Brand Reputation Mining and Visualisation
Abstract
Online brand reputation is of increasing significance to many organisations and institutes around the globe. As the usage of the www continues to increase it has become the most commonly used platform for users and customers of services and products to discuss their views and experiences. The nature of this www discussion can significantly influence the perception and hence the success of a brand. Brand Reputation Mining (BRM) is a process to help brand owners to know what is being said about their brand online. This paper proposes a BRM framework to provide support for enterprises wishing to conduct brand reputation management. The proposed framework can be generically applied to collect, process and display the reputation of different brands. A key feature is the visualisation facilities included to allow the display of the results of reputation mining activities. The framework is fully described and illustrated using a case study. The concepts expressed in this paper have been incorporated into the “LittleBirdy” brand reputation management product commercially available from Hit Search Ltd.
Ayesh Alshukri, Frans Coenen, Yang Li, Andrew Redfern, Prudence W. H. Wong
A Review of Voice Activity Detection Techniques for On-Device Isolated Digit Recognition on Mobile Devices
Abstract
This paper presents a review of different Voice Activity Detection (VAD) techniques that can be easily applied to On-device Isolated digit recognition on a mobile device. Techniques investigated include; Short Time Energy, Linear predictive coding residual (prediction error), Discrete Fourier Transform (DFT) based linear cross correlation and K-means clustering based VAD. The optimum VAD technique was found to be K-means clustering of Prediction error which gives a recognition rate of 86.6 %. This technique will be further used with an LPC based speech recognition algorithm for digit recognition on the mobile device.
M. K. Mustafa, Tony Allen, Lindsay Evett

Short Papers

Frontmatter
Ontology-Based Information Extraction and Reservoir Computing for Topic Detection from Blogosphere’s Content: A Case Study About BBC Backstage
Abstract
This research study aims at detecting topics and extracting themes (subtopics) from the blogosphere’s content while bridging the gap between the Social Web and the Semantic Web. The goal is to detect certain types of information from collections of blogs’ and microblogs’ narratives that lack explicit semantics. The approach presented introduces a novel approach that blends together two young paradigms: Ontology-Based Information Extraction (OBIE) and Reservoir Computing (RC). The novelty of the work lies in integrating ontologies and RC as well as the pioneering use of RC with social media data. Experiments with retrospect data from blogs and Twitter microblogs provide valuable insights into the BBC Backstage initiative and prove the viability of the approach presented in terms of scalability, computational complexity, and performance.
M. Arguello-Casteleiro, M. J. Fernandez-Prieto
A Study on Road Junction Control Method Selection Using an Artificial Intelligent Multi-criteria Decision Making Framework
Abstract
With the increasing number of vehicles on roads, choosing a proper Road Junction Control (RJC) Method has become an important decision for reducing traffic congestion and cost. However, the public awareness of environmental sustainability and diverse voices from different stakeholders make such decision a knotty one. In this paper, an artificial intelligent decision-making framework using Hierarchical Half Fuzzy TOPSIS (HHF-TOPSIS) is proposed for RJC method selection. Compared with the existing qualitative comparison method suggested in the Design Manual for Roads and Bridges, this method can provide a more efficient and objective approach to reach the best compromise against all relevant objectives.
P. K. Kwok, D. W. H. Chau, H. Y. K. Lau
Metadaten
Titel
Research and Development in Intelligent Systems XXXI
herausgegeben von
Max Bramer
Miltos Petridis
Copyright-Jahr
2014
Electronic ISBN
978-3-319-12069-0
Print ISBN
978-3-319-12068-3
DOI
https://doi.org/10.1007/978-3-319-12069-0