Skip to main content
Top

2014 | Book

Artificial Neural Networks and Machine Learning – ICANN 2014

24th International Conference on Artificial Neural Networks, Hamburg, Germany, September 15-19, 2014. Proceedings

Editors: Stefan Wermter, Cornelius Weber, Włodzisław Duch, Timo Honkela, Petia Koprinkova-Hristova, Sven Magg, Günther Palm, Alessandro E. P. Villa

Publisher: Springer International Publishing

Book Series : Lecture Notes in Computer Science

insite
SEARCH

About this book

The book constitutes the proceedings of the 24th International Conference on Artificial Neural Networks, ICANN 2014, held in Hamburg, Germany, in September 2014.

The 107 papers included in the proceedings were carefully reviewed and selected from 173 submissions. The focus of the papers is on following topics: recurrent networks; competitive learning and self-organisation; clustering and classification; trees and graphs; human-machine interaction; deep networks; theory; reinforcement learning and action; vision; supervised learning; dynamical models and time series; neuroscience; and applications.

Table of Contents

Frontmatter

Recurrent Networks

Sequence Learning

Dynamic Cortex Memory: Enhancing Recurrent Neural Networks for Gradient-Based Sequence Learning

In this paper a novel recurrent neural network (RNN) model for gradient-based sequence learning is introduced. The presented dynamic cortex memory (DCM) is an extension of the well-known long short term memory (LSTM) model. The main innovation of the DCM is the enhancement of the inner interplay of the gates and the error carousel due to several new and trainable connections. These connections enable a direct signal transfer from the gates to one another. With this novel enhancement the networks are able to converge faster during training with back-propagation through time (BPTT) than LSTM under the same training conditions. Furthermore, DCMs yield better generalization results than LSTMs. This behaviour is shown for different supervised problem scenarios, including storing precise values, adding and learning a context-sensitive grammar.

Sebastian Otte, Marcus Liwicki, Andreas Zell
Learning and Recognition of Multiple Fluctuating Temporal Patterns Using S-CTRNN

In the present study, we demonstrate the learning and recognition capabilities of our recently proposed recurrent neural network (RNN) model called stochastic continuous-time RNN (S-CTRNN). S-CTRNN can learn to predict not only the mean but also the variance of the next state of the learning targets. The network parameters consisting of weights, biases, and initial states of context neurons are optimized through maximum likelihood estimation (MLE) using the gradient descent method. First, we clarify the essential difference between the learning capabilities of conventional CTRNN and S-CTRNN by analyzing the results of a numerical experiment in which multiple fluctuating temporal patterns were used as training data, where the variance of the Gaussian noise varied among the patterns. Furthermore, we also show that the trained S-CTRNN can recognize given fluctuating patterns by inferring the initial states that can reproduce the patterns through the same MLE scheme as that used for network training.

Shingo Murata, Hiroaki Arie, Tetsuya Ogata, Jun Tani, Shigeki Sugano
Regularized Recurrent Neural Networks for Data Efficient Dual-Task Learning

We introduce a regularization technique to improve system identification for dual-task learning with recurrent neural networks. In particular, the method is introduced using the Factored Tensor Recurrent Neural Networks first presented in [1]. Our goal is to identify a dynamical system with few available observations by augmenting them with data from a sufficiently observed similar system. In our previous work, we discovered that the model accuracy degrades whenever little data of the system of interest is available. The presented regularization term in this work allows to significantly reduce the model error thereby improving the exploitation of knowledge of the well observed system. This scenario is crucial in many real world applications, where data efficiency plays an important role. We motivate the problem setting and our regularized dual-task learning approach by industrial use cases, e.g. gas or wind turbine modeling for optimization and monitoring. Then, we formalize the problem and describe our regularization term by which the learning objective of the Factored Tensor Recurrent Neural Network is extended. Finally, we demonstrate its effectiveness on the cart-pole and mountain car benchmarks.

Sigurd Spieckermann, Siegmund Düll, Steffen Udluft, Thomas Runkler

Echo State Networks

On-line Training of ESN and IP Tuning Effect

In the present paper we investigate influence of IP tuning of Echo state network (ESN) reservoir on the overall behavior of the on-line trained adaptive critic network. The experiments were done using Adaptive Critic Design (ACD) scheme with on-line trainable ESN critic for real time control of a mobile laboratory robot. Comparison of behavior of ESN critics trained with and without IP tuning showed that IP algorithm improved critic behavior significantly. It was observed that IP tuning prevents uncontrolled increase of reservoir output weights during on-line training.

Petia Koprinkova-Hristova
An Incremental Approach to Language Acquisition: Thematic Role Assignment with Echo State Networks

In previous research a model for thematic role assignment (

θ

RARes) was proposed, using the Reservoir Computing paradigm. This language comprehension model consisted of a recurrent neural network (RNN) with fixed random connections which models distributed processing in the prefrontal cortex, and an output layer which models the striatum. In contrast to this previous batch learning method, in this paper we explored a more biological learning mechanism. A new version of the model (i-

θ

RARes) was developed that permitted incremental learning, at each time step. Learning was based on a stochastic gradient descent method. We report here results showing that this incremental version was successfully able to learn a corpus of complex grammatical constructions, reinforcing the neurocognitive plausibility of the model from a language acquisition perspective.

Xavier Hinaut, Stefan Wermter
Memory Capacity of Input-Driven Echo State Networks at the Edge of Chaos

Reservoir computing provides a promising approach to efficient training of recurrent neural networks, by exploiting the computational properties of the reservoir structure. Various approaches, ranging from suitable initialization to reservoir optimization by training have been proposed. In this paper we take a closer look at short-term memory capacity, introduced by Jaeger in case of echo state networks. Memory capacity has recently been investigated with respect to criticality, the so called edge of chaos, when the network switches from a stable regime to an unstable dynamic regime. We calculate memory capacity of the networks for various input data sets, both random and structured, and show how the data distribution affects the network performance. We also investigate the effect of reservoir sparsity in this context.

Peter Barančok, Igor Farkaš
Adaptive Critical Reservoirs with Power Law Forgetting of Unexpected Input Sequences

The echo-state condition names an upper limit for the hidden layer connectivity in recurrent neural networks. If the network is below this limit there is an injective, continuous mapping from the recent input history to the internal state of the network. Above the network becomes chaotic, the dependence on the initial state of the network may never be washed out. I focus on the biological relevance of echo state networks with a critical connectivity strength at the separation line between these two conditions and discuss some related biological findings, i.e. there is evidence that the neural connectivity in cortical slices is tuned to a critical level. In addition, I propose a model that makes use of a special learning mechanism within the recurrent layer and the input connectivity. Results show that after adaptation indeed traces of single unexpected events stay for a longer time period than exponential in the network.

Norbert Michael Mayer

Recurrent Network Theory

Interactive Evolving Recurrent Neural Networks Are Super-Turing Universal

Understanding the dynamical and computational capabilities of neural models represents an issue of central importance. In this context, recent results show that interactive evolving recurrent neural networks are super-Turing, irrespective of whether their synaptic weights are rational or real. We extend these results by showing that interactive evolving recurrent neural networks are not only super-Turing, but also capable of simulating any other possible interactive deterministic system. In this sense, interactive evolving recurrent neural networks represents a super-Turing universal model of computation, irrespective of whether their synaptic weights are rational or real.

Jérémie Cabessa, Alessandro E. P. Villa
Attractor Metadynamics in Adapting Neural Networks

Slow adaption processes, like synaptic and intrinsic plasticity, abound in the brain and shape the landscape for the neural dynamics occurring on substantially faster timescales. At any given time the network is characterized by a set of internal parameters, which are adapting continuously, albeit slowly. This set of parameters defines the number and the location of the respective adiabatic attractors. The slow evolution of network parameters hence induces an evolving attractor landscape, a process which we term attractor metadynamics. We study the nature of the metadynamics of the attractor landscape for several continuous-time autonomous model networks. We find both first- and second-order changes in the location of adiabatic attractors and argue that the study of the continuously evolving attractor landscape constitutes a powerful tool for understanding the overall development of the neural dynamics.

Claudius Gros, Mathias Linkerhand, Valentin Walther
Basic Feature Quantities of Digital Spike Maps

The digital spike-phase map is a simple digital dynamical system that can generate various spike-trains. In order to approach systematic analysis of the steady and transient states, four basic feature quantities are presented. Using the quantities, we analyze an example based on the bifurcating neuron with triangular base signal and consider basic four cases of the spike-train dynamics.

Hiroki Yamaoka, Narutoshi Horimoto, Toshimichi Saito

Competitive Learning and Self-Organisation

Discriminative Fast Soft Competitive Learning

Proximity matrices like kernels or dissimilarity matrices provide non-standard data representations common in the life science domain. Here we extend fast soft competitive learning to a discriminative and vector labeled learning algorithm for proximity data. It provides a more stable and consistent integration of label information in the cost function solely based on a give proximity matrix without the need of an explicite vector space. The algorithm has linear computational and memory requirements and performs favorable to traditional techniques.

Frank-Michael Schleif
Human Action Recognition with Hierarchical Growing Neural Gas Learning

We propose a novel biologically inspired framework for the recognition of human full-body actions. First, we extract body pose and motion features from depth map sequences. We then cluster pose-motion cues with a two-stream hierarchical architecture based on growing neural gas (GNG). Multi-cue trajectories are finally combined to provide prototypical action dynamics in the joint feature space. We extend the unsupervised GNG with two labelling functions for classifying clustered trajectories. Noisy samples are automatically detected and removed from the training and the testing set. Experiments on a set of 10 human actions show that the use of multi-cue learning leads to substantially increased recognition accuracy over the single-cue approach and the learning of joint pose-motion vectors.

German Ignacio Parisi, Cornelius Weber, Stefan Wermter
Real-Time Anomaly Detection with a Growing Neural Gas

We present a novel system for vision based anomaly detection in real-time environments. Our system uses an event-based vision sensor consisting of asynchronously operating pixels that is inspired by the human retina. Each pixel reports events of illumination changes, are processed in a purely event-based tracker that pursues edges of events in the input stream. The tracker estimates are used to determine whether the input events originate from anomalous or regular data. We distinguish between the two cases with a Growing Neural Gas (GNG), which is modified to suite our event-based processing pipeline. While learning of the GNG is supervised and performed offline, the detection is carried out online. We evaluate our system by inspection of fast-spinning cog-wheels. Our system achieves faster than real-time speed on commodity hardware and generalizes well to other cases. The results of this paper can be applied both to technical implementations where high speed but little processing power is required, and for further investigations into event-based algorithms.

Nicolai Waniek, Simon Bremer, Jörg Conradt
Classification with Reject Option Using the Self-Organizing Map

Reject option is a technique used to improve classifier’s reliability in decision support systems. It consists on withholding the automatic classification of an item, if the decision is considered not sufficiently reliable. The rejected item is then handled by a different classifier or by a human expert. The vast majority of the works on this issue have been concerned with implementing a reject option by endowing a supervised learning scheme (e.g., Multilayer Perceptron, Learning Vector Quantization or Support Vector Machines) with a reject mechanism. In this paper we introduce variants of the Self-Organizing Map (SOM), originally an unsupervised learning scheme, to act as supervised classifiers with reject option, and compare their performances with that of the MLP classifier.

Ricardo Sousa, Ajalmar R. da Rocha Neto, Jaime S. Cardoso, Guilherme A. Barreto

Clustering and Classification

A Non-parametric Maximum Entropy Clustering

Clustering is a fundamental tool for exploratory data analysis. Information theoretic clustering is based on the optimization of information theoretic quantities such as entropy and mutual information. Recently, since these quantities can be estimated in non-parametric manner, non-parametric information theoretic clustering gains much attention. Assuming the dataset is sampled from a certain cluster, and assigning different sampling weights depending on the clusters, the cluster conditional information theoretic quantities are estimated. In this paper, a simple clustering algorithm is proposed based on the principle of maximum entropy. The algorithm is experimentally shown to be comparable to or outperform conventional non-parametric clustering methods.

Hideitsu Hino, Noboru Murata
Instance Selection Using Two Phase Collaborative Neighbor Representation

Finding relevant instances in databases has always been a challenging task. Recently a new method, called

Sparse Modeling Representative Selection

(SMRS) has been proposed in this area and is based on data self-representation. SMRS estimates a matrix of coefficients by minimizing a reconstruction error and a regularization term on these coefficients using the

L

1,

q

matrix norm. In this paper, we propose another alternative of coding based on a two stage Collaborative Neighbor Representation in which a non-dense matrix of coefficients is estimated without invoking any explicit sparse coding. Experiments are conducted on summarizing a video movie and on summarizing training face datasets used for face recognition. These experiments showed that the proposed method can outperform the state-of-the art methods.

Fadi Dornaika, I. Kamal Aldine
Global Metric Learning by Gradient Descent

The

k

-NN classifier can be very competitive if an appropriate distance measure is used. It is often used in applications because the classification decisions are easy to interpret. Here, we demonstrate how to find a good Mahalanobis distance for

k

-NN classification by a simple gradient descent without any constraints. The cost term uses global distances and unlike other methods there is a soft transition in the influence of data points. It is evaluated and compared to other metric learning and feature weighting methods on datasets from the UCI repository, where the described gradient method also shows a high robustness. In the comparison the advantages of global approaches are demonstrated.

Jens Hocke, Thomas Martinetz
Leaving Local Optima in Unsupervised Kernel Regression

Embedding high-dimensional patterns in low-dimensional latent spaces is a challenging task. In this paper, we introduce re-sampling strategies to leave local optima in the data space reconstruction error (DSRE) minimization process of unsupervised kernel regression (UKR). For this sake, we concentrate on a hybrid UKR variant that combines iterative solution construction with gradient descent based optimization. Patterns with high reconstruction errors are removed from the manifold and re-sampled based on Gaussian sampling. Re-sampling variants consider different pattern reconstruction errors, varying numbers of re-sampled patterns, and termination conditions. The re-sampling process with UKR can also improve ISOMAP embeddings. Experiments on typical benchmark data sets illustrate the capabilities of strategies for leaving optima.

Daniel Lückehe, Oliver Kramer

Trees and Graphs

An Algorithm for Directed Graph Estimation

A problem of estimating the intrinsic graph structure from observed data is considered. The observed data in this study is a matrix with elements representing

dependency

between nodes in the graph. Each element of the observed matrix represents, for example, co-occurrence of events at two nodes, or correlation of variables corresponding to two nodes. The dependency does not represent direct connections and includes influences of various paths, and spurious correlations make the estimation of direct connection difficult. To alleviate this difficulty, digraph Laplacian is used for characterizing a graph. A generative model of an observed matrix is proposed, and a parameter estimation algorithm for the model is also proposed. The proposed method is capable of dealing with directed graphs, while conventional graph structure estimation methods from an observed matrix are only applicable to undirected graphs. Experimental result shows that the proposed algorithm is able to identify the intrinsic graph structure.

Hideitsu Hino, Atsushi Noda, Masami Tatsuno, Shotaro Akaho, Noboru Murata
Merging Strategy for Local Model Networks Based on the Lolimot Algorithm

In this paper an extension of the established training algorithm for nonlinear system identification called

Lolimot

is presented [9]. It is a heuristic tree-construction method that trains a local linear neuro-fuzzy network. Due to its very simple partitioning strategy,

Lolimot

is a fast and robust modeling approach, but has a limited flexibility. Therefore a new merging approach for regression tasks is presented, that can rearrange the local model structure in the input space, without harming the global model complexity.

Torsten Fischer, Oliver Nelles
Factor Graph Inference Engine on the SpiNNaker Neural Computing System

This paper presents a novel method for implementing Factor Graphs in a SpiNNaker neural computing system. The SpiNNaker system provides resources for fine-grained parallelism, designed for implementing a distributed computing system. We present a framework which utilizes available SpiNNaker resources to implement a discrete Factor Graph: a powerful graphical model for probabilistic inference. Our framework allows mapping and routing a Factor Graph on the SpiNNaker hardware using SpiNNaker’s event-based communication system. An example application of the proposed framework in a real-world robotics scenario is given and the result shows that the framework can handle computation of 26.14 MFLOPS only in 30.5ms. We demonstrate that the framework easily extends for larger Factor Graph networks in a bigger SpiNNaker system, which makes it suitable for complex and challenging computational intelligence tasks.

Indar Sugiarto, Jörg Conradt
High-Dimensional Binary Pattern Classification by Scalar Neural Network Tree

The paper offers an algorithm (SNN-tree) that extends the binary tree search algorithm so that it can deal with distorted input vectors. Perceptrons are the tree nodes. The algorithm features an iterative solution search and stopping criterion. Unlike the SNN-tree algorithm, popular methods (LSH, k-d tree, BBF-tree, spill-tree) stop working as the dimensionality of the space grows (

N

> 1000). With such high dimensionality, our algorithm works 7 times faster than the exhaustive search algorithm.

Vladimir Kryzhanovsky, Magomed Malsagov, Juan Antonio Clares Tomas, Irina Zhelavskaya

Human-Machine Interaction

Human Activity Recognition on Smartphones with Awareness of Basic Activities and Postural Transitions

Postural Transitions (PTs) are transitory movements that describe the change of state from one static posture to another. In several Human Activity Recognition (HAR) systems, these transitions cannot be disregarded due to their noticeable incidence with respect to the duration of other Basic Activities (BAs). In this work, we propose an online smartphone-based HAR system which deals with the occurrence of postural transitions. If treated properly, the system accuracy improves by avoiding fluctuations in the classifier. The method consists of concurrently exploiting Support Vector Machines (SVMs) and temporal filters of activity probability estimations within a limited time window. We present the benefits of this approach through experiments over a HAR dataset which has been updated with PTs and made publicly available. We also show the new approach performs better than a previous baseline system, where PTs were not taken into account.

Jorge-Luis Reyes-Ortiz, Luca Oneto, Alessandro Ghio, Albert Samá, Davide Anguita, Xavier Parra
sNN-LDS: Spatio-temporal Non-negative Sparse Coding for Human Action Recognition

Current state-of-the-art approaches for visual human action recognition focus on complex local spatio-temporal descriptors, while the spatio-temporal relations between the descriptors are discarded. These bag-of-features (BOF) based methods come with the disadvantage of limited descriptive power, because class-specific mid- and large-scale spatio-temporal information, such as body pose sequences, cannot be represented. To overcome this restriction, we propose sparse non-negative linear dynamical systems (sNN-LDS) as a dynamic, parts-based, spatio-temporal representation of local descriptors. We provide novel learning rules based on sparse non-negative matrix factorization (sNMF) to simultaneously learn both the parts as well as their transitions. On the challenging UCF-Sports dataset our sNN-LDS combined with simple local features is competitive with state-of-the-art BOF-SVM methods.

Thomas Guthier, Adrian Šošić, Volker Willert, Julian Eggert
Interactive Language Understanding with Multiple Timescale Recurrent Neural Networks

Natural language processing in the human brain is complex and dynamic. Models for understanding, how the brain’s architecture acquires language, need to take into account the temporal dynamics of verbal utterances as well as of action and visual embodied perception. We propose an architecture based on three Multiple Timescale Recurrent Neural Networks (MTRNNs) interlinked in a cell assembly that learns verbal utterances grounded in dynamic proprioceptive and visual information. Results show that the architecture is able to describe novel dynamic actions with correct novel utterances, and they also indicate that multi-modal integration allows for a disambiguation of concepts.

Stefan Heinrich, Stefan Wermter
A Neural Dynamic Architecture Resolves Phrases about Spatial Relations in Visual Scenes

How spatial language, important to both cognitive science and robotics, is mapped to real-world scenes by neural processes is not understood. We present an autonomous neural dynamics that achieves this mapping flexibly. Neural activation fields represent and spatially transform perceptual information. An architecture of dynamic nodes interacts with these perceptual fields to instantiate categorical concepts. Discrete time processing steps emerge from instabilities of the time-continuous neural dynamics and are organized sequentially by these nodes. These steps include the attentional selection of individual objects in a scene, mapping locations to an object-centered reference frame, and evaluating matches to relational spatial terms. The architecture can respond to queries specified by setting the state of discrete nodes. It autonomously generates a response based on visual input about a scene.

Mathis Richter, Jonas Lins, Sebastian Schneegans, Gregor Schöner
Chinese Image Character Recognition Using DNN and Machine Simulated Training Samples

Inspired by the success of deep neural network (DNN) models in solving challenging visual problems, this paper studies the task of Chinese Image Character Recognition (ChnICR) by leveraging DNN model and huge machine simulated training samples. To generate the samples, clean machine born Chinese characters are extracted and are plus with common variations of image characters such as changes in size, font, boldness, shift and complex backgrounds, which in total produces over 28 million character images, covering the vast majority of occurrences of Chinese character in real life images. Based on these samples, a DNN training procedure is employed to learn the appropriate Chinese character recognizer, where the width and depth of DNN, and the volume of samples are empirically discussed. Parallel to this, a holistic Chinese image text recognition system is developed. Encouraging experimental results on text from 13 TV channels demonstrate the effectiveness of the learned recognizer, from which significant performance gains are observed compared to the baseline system.

Jinfeng Bai, Zhineng Chen, Bailan Feng, Bo Xu
Polyphonic Music Generation by Modeling Temporal Dependencies Using a RNN-DBN

In this paper, we propose a generic technique to model temporal dependencies and sequences using a combination of a recurrent neural network and a Deep Belief Network. Our technique, RNN-DBN, is an amalgamation of the memory state of the RNN that allows it to provide temporal information and a multi-layer DBN that helps in high level representation of the data. This makes RNN-DBNs ideal for sequence generation. Further, the use of a DBN in conjunction with the RNN makes this model capable of significantly more complex data representation than a Restricted Boltzmann Machine (RBM). We apply this technique to the task of polyphonic music generation.

Kratarth Goel, Raunaq Vohra, J. K. Sahoo
On Improving the Classification Capability of Reservoir Computing for Arabic Speech Recognition

Designing noise-resilient systems is a major challenge in the field of automated speech recognition (ASR). These systems are crucial for real-world applications where high levels of noise tend to be present. We introduce a noise robust system based on Echo State Networks and Extreme Kernel machines which we call ESNEKM. To evaluate the performance of the proposed system, we used our recently released public Arabic speech dataset and the well-known spoken Arabic digits (SAD) dataset. Different feature extraction methods considered in this study include mel-frequency cepstral coefficients (MFCCs), perceptual linear prediction (PLP) and RASTA- perceptual linear prediction. These extracted features were fed to the ESNEKM and the result compared with a baseline hidden Markov model (HMM), so that nine models were compared in total. ESNEKM models outperformed HMM models under all the feature extraction methods, noise levels, and noise types. The best performance was obtained by the model that combined RASTA-PLP with ESNEKM.

Abdulrahman Alalshekmubarak, Leslie S. Smith
Neural Network Based Data Fusion for Hand Pose Recognition with Multiple ToF Sensors

We present a study on 3D based hand pose recognition using a new generation of low-cost time-of-flight(ToF) sensors intended for outdoor use in automotive human-machine interaction. As signal quality is impaired compared to Kinect-type sensors, we study several ways to improve performance when a large number of gesture classes is involved. We investigate the performance of different 3D descriptors, as well as the fusion of two ToF sensor streams. By basing a data fusion strategy on the fact that multilayer perceptrons can produce normalized confidences individually for each class, and similarly by designing information-theoretic online measures for assessing confidences of decisions, we show that appropriately chosen fusion strategies can improve overall performance to a very satisfactory level. Real-time capability is retained as the used 3D descriptors, the fusion strategy as well as the online confidence measures are computationally efficient.

Thomas Kopinski, Alexander Gepperth, Stefan Geisler, Uwe Handmann
Sparse Single-Hidden Layer Feedforward Network for Mapping Natural Language Questions to SQL Queries

Mapping natural language (NL) statements into SQL queries allows users to interact with systems through everyday language. Semantic parsing has seen a growing interest over the past decades. In this paper, we extend single hidden layer feedforward network (SLFN) by adding the Kullback-Liebler (KL) divergence parameter to its objective function. We refer to this algorithm as Sparse SLFN (S-SLFN) which can learn whether an SQL query answers a particular NL question. With Bag of Words (BoW) representing the questions and the queries, the algorithm, by enforcing sparsity, is meant to retain robust features representing informative relationships and structure of the data. Experimental results show that S-SLFN outperforms SLFN and other algorithms for the GeoQueries dataset by a respectable margin.

Issam H. Laradji, Lahouari Ghouti, Faisal Saleh, Musab A. AlTurki
Towards Context-Dependence Eye Movements Prediction in Smart Meeting Rooms

Being able to predict gaze locations, as compared to only measuring them, is desirable in many systems such as the design of web pages and commercials adaptive user interfaces, interactive visualization, or attention management systems. However, accurately predicting eye movements remains a challenging problem. In this paper, we present the results of experimental study to improve the prediction of saliency maps in smart meeting rooms. More specifically, we investigate meeting scenarios in terms of their context-dependence saliency based on different image features. We have recorded the center of gaze of users in meeting rooms in different scenarios (giving a talk, listening). We then used a data-driven approach to find out which features are important in each scenario. We found that the predictions differ according to the type of features we selected. Most interestingly, we found that models trained on face features perform better than the models trained on other features in the giving a talk scenario, but in the listening scenario the models trained on competing saliency features from Itti and Koch perform better than the models trained on another features. This finding points towards including context information about the scene and situation into the computation of saliency maps as important towards developing models of eye movements, which operate well under natural conditions such as those encountered in ubiquitous computing settings.

Redwan Abdo A. Mohammed, Lars Schwabe, Oliver Staadt

Deep Networks

Variational EM Learning of DSBNs with Conditional Deep Boltzmann Machines

Variational EM (VEM) is an efficient parameter learning scheme for sigmoid belief networks with many layers of latent variables. The choice of the inference model that forms the variational lower bound of the log likelihood is critical in VEM learning. The mean field approximations and wake-sleep algorithm use simple models that are computationally efficient, but may be poor approximations to the true posterior densities when the latent variables have strong mutual dependencies. In this paper, we describe a variational EM learning method of DSBNs with a new inference model known as the

conditional deep Boltzmann machine

(cDBM), which is an

undirected

graphical model capable of representing complex dependencies among latent variables. We show that this algorithm does not require the computation of the intractable partition function in the undirected cDBM model, and can be accelerated with contrastive learning. Performances of the proposed method are evaluated and compared on handwritten digit data.

Xing Zhang, Siwei Lyu
Improving Deep Neural Network Performance by Reusing Features Trained with Transductive Transference

Transfer Learning is a paradigm in machine learning to solve a target problem by reusing the learning with minor modifications from a different but related source problem. In this paper we propose a novel feature transference approach, especially when the source and the target problems are drawn from different distributions. We use deep neural networks to transfer either low or middle or higher-layer features for a machine trained in either unsupervised or supervised way. Applying this feature transference approach on Convolutional Neural Network and Stacked Denoising Autoencoder on four different datasets, we achieve lower classification error rate with significant reduction in computation time with lower-layer features trained in supervised way and higher-layer features trained in unsupervised way for classifying images of uppercase and lowercase letters dataset.

Chetak Kandaswamy, Luís M. Silva, Luís A. Alexandre, Jorge M. Santos, Joaquim Marques de Sá
From Maxout to Channel-Out: Encoding Information on Sparse Pathways

Motivated by an important insight from neural science that “functionality is determined by pathway”, we propose a new deep network framework, called “channel-out network”, which encodes information on sparse pathways. We argue that the recent success of maxout networks can also be explained by its ability of encoding information on sparse pathways, while channel-out network does not only select pathways at training time but also at inference time. From a mathematical perspective, channel-out networks can represent a wider class of piece-wise continuous functions, thereby endowing the network with more expressive power than that of maxout networks. We test our channel-out networks on several well-known image classification benchmarks, achieving new state-of-the-art performances on CIFAR-100 and STL-10.

Qi Wang, Joseph JaJa
Minimizing Computation in Convolutional Neural Networks

Convolutional Neural Networks (CNNs) have been successfully used for many computer vision applications. It would be beneficial to these applications if the computational workload of CNNs could be reduced. In this work we analyze the linear algebraic properties of CNNs and propose an algorithmic modification to reduce their computational workload. An up to a 47% reduction can be achieved without any change in the image recognition results or the addition of any hardware accelerators.

Jason Cong, Bingjun Xiao
One-Shot Learning with Feedback for Multi-layered Convolutional Network

This paper proposes an improved add-if-silent rule, which is suited for training intermediate layers of a multi-layered convolutional network, such as a neocognitron. By the add-if-silent rule, a new cell is generated if all postsynaptic cells are silent. The generated cell learns the activity of the presynaptic cells in one-shot, and its input connections will never be modified afterward. To use this learning rule for a convolutional network, it is required to decide at which retinotopic location this rule is to be applied. In the conventional add-if-silent rule, we chose the location where the activity of presynaptic cells is the largest. In the proposed new learning rule, a negative feedback is introduced from postsynaptic cells to presynaptic cells, and a new cell is generated at the location where the presynaptic activity fails to be suppressed by the feedback. We apply this learning rule to a neocognitron for hand-written digit recognition, and demonstrate the decrease in the recognition error.

Kunihiko Fukushima

Theory

Optimization

Row-Action Projections for Nonnegative Matrix Factorization

Nonnegative Matrix Factorization (NMF) is more and more frequently used for analyzing large-scale nonnegative data, where the number of samples and/or the number of observed variables is large. In the paper, we discuss two applications of the row-action projections in the context of learning latent factors from large-scale data. First, we show that they can be efficiently used for improving the on-line learning in dynamic NMF. Next, they can also considerably reduce the computational complexity of the optimization algorithms used for factor learning from strongly redundant data. The experiments demonstrate high efficiency of the proposed methods.

Rafał Zdunek
Structure Perturbation Optimization for Hopfield-Type Neural Networks

In this paper, we extract the core idea of state perturbation from

Hopfield-type

neural networks and define state perturbation formulas to describe the general way of optimization methods. Departing from the core idea and the formulas, we propose a novel optimization method related to neural network structure, named structure perturbation optimization. Our method can produce a structure transforming process to retrain

Hopfield-type

neural networks to get better problem-solving ability. Experiments validate that our method effectively helps

Hopfield-type

neural networks to escape from local minima and get superior solutions.

Gang Yang, Xirong Li, Jieping Xu, Qin Jin
Complex-Valued Multilayer Perceptron Search Utilizing Singular Regions of Complex-Valued Parameter Space

In the search space of a complex-valued multilayer perceptron having

J

hidden units, C-MLP(

J

), there exist flat areas called singular regions, as is the case with a real-valued MLP. The singular regions cause serious stagnation of learning, preventing usual search methods from finding an excellent solution. However, there exist descending paths from the regions since most points in the regions are saddles. This paper proposes a completely new learning method that does not avoid but makes good use of singular regions to stably and successively find excellent solutions commensurate with C-MLP(

J

). Our experiments showed the proposed method worked well.

Seiya Satoh, Ryohei Nakano

Layered Networks

Mix-Matrix Transformation Method for Max-Сut Problem

One usually tries to raise the efficiency of optimization techniques by changing the dynamics of local optimization. In contrast to the above approach, we propose changing the surface of the problem rather than the dynamics of local search. The Mix-Matrix algorithm proposed by the authors previously [1] realizes such transformation and can be applied directly to a max-cut problem and successfully compete with other popular algorithms in this field such as CirCut and Scatter Search.

Iakov Karandashev, Boris Kryzhanovsky
Complexity of Shallow Networks Representing Functions with Large Variations

Model complexities of networks representing multivariable functions is studied in terms of variational norms tailored to types of network units. It is shown that the size of the variational norm reflects both the number of hidden units and sizes of output weights. Lower bounds on growth of variational norms with increasing input dimension

d

are derived for Gaussian units and perceptrons. It is proven that variation of the

d

-dimensional parity with respect to Gaussian Support Vector Machines grows exponentially with

d

and for large values of

d

, almost any randomly-chosen Boolean function has variation with respect to perceptrons depending on

d

exponentially.

Věra Kůrková, Marcello Sanguineti
Visualizing Hierarchical Representation in a Multilayered Restricted RBF Network

In this study we propose a hierarchical neural network that is able to generate a topographical map in its internal layer. The map significantly differs from the conventional Kohonen’s SOM, in that it preserves the topological characteristics in relevance to the context, for example the labels, of the data. This map is useful if we are interested in visualizing the underlying characteristics of the classificability of the data that traditionally cannot be visualized with the standard SOM. In this paper, we expand our network into a multilayered structure that allows us visualize and thus better understand on how the neural network perceives the given data in the light of classification task.

Pitoyo Hartono, Paul Hollensen, Thomas Trappenberg

Reinforcement Learning and Action

Contingent Features for Reinforcement Learning

Applying reinforcement learning algorithms in real-world domains is challenging because relevant state information is often embedded in a stream of high-dimensional sensor data. This paper describes a novel algorithm for learning task-relevant features through interactions with the environment. The key idea is that a feature is likely to be useful to the degree that its dynamics can be controlled by the actions of the agent. We describe an algorithm that can find such features and we demonstrate its effectiveness in an artificial domain.

Nathan Sprague
A Non-stationary Infinite Partially-Observable Markov Decision Process

Partially Observable Markov Decision Processes (POMDPs) have been met with great success in planning domains where agents must balance actions that provide knowledge and actions that provide reward. Recently, nonparametric Bayesian methods have been successfully applied to POMDPs to obviate the need of a priori knowledge of the size of the state space, allowing to assume that the number of visited states may grow as the agent explores its environment. These approaches rely on the assumption that the agent’s environment remains stationary; however, in real-world scenarios the environment may change over time. In this work, we aim to address this inadequacy by introducing a dynamic nonparametric Bayesian POMDP model that both allows for automatic inference of the (distributional) representations of POMDP states, and for capturing non-stationarity in the modeled environments. Formulation of our method is based on imposition of a suitable dynamic hierarchical Dirichlet process (dHDP) prior over state transitions. We derive efficient algorithms for model inference and action planning and evaluate it on several benchmark tasks.

Sotirios P. Chatzis, Dimitrios Kosmopoulos
Tool-Body Assimilation Model Based on Body Babbling and a Neuro-Dynamical System for Motion Generation

We propose a model for robots to use tools without predetermined parameters based on a human cognitive model. Almost all existing studies of robot using tool require predetermined motions and tool features, so the motion patterns are limited and the robots cannot use new tools. Other studies use a full search for new tools; however, this entails an enormous number of calculations. We built a model for tool use based on the phenomenon of tool-body assimilation using the following approach: We used a humanoid robot model to generate random motion, based on human body babbling. These rich motion experiences were then used to train a recurrent neural network for modeling a body image. Tool features were self-organized in the parametric bias modulating the body image according to the used tool. Finally, we designed the neural network for the robot to generate motion only from the target image.

Kuniyuki Takahashi, Tetsuya Ogata, Hadi Tjandra, Shingo Murata, Hiroaki Arie, Shigeki Sugano
A Gaussian Process Reinforcement Learning Algorithm with Adaptability and Minimal Tuning Requirements

We present a novel Bayesian reinforcement learning algorithm that addresses model bias and exploration overhead issues. The algorithm combines different aspects of several state-of-the-art reinforcement learning methods that use Gaussian Processes model-based approaches to increase the use of the online data samples. The algorithm uses a smooth reward function requiring the reward value to be derived from the environment state. It works with continuous states and actions in a coherent way with a minimized need for expert knowledge in parameter tuning. We analyse and discuss the practical benefits of the selected approach in comparison to more traditional methodological choices, and illustrate the use of the algorithm in a motor control problem involving a two-link simulated arm.

Jonathan Strahl, Timo Honkela, Paul Wagner
Sensorimotor Control Learning Using a New Adaptive Spiking Neuro-Fuzzy Machine, Spike-IDS and STDP

Human mind from system perspective deals with high dimensional complex world as an adaptive Multi-Input Multi-Output complex system. This view is theorized by reductionism theory in philosophy of mind, where the world is represented as logical combination of simpler sub-systems for human so that operate with less energy. On the other hand, Human usually uses linguistic rules to describe and manipulate his expert knowledge about the world; the way that is well modeled by Fuzzy Logic. But how such a symbolic form of knowledge can be encoded and stored in plausible neural circuitry? Based on mentioned postulates, we have proposed an adaptive Neuro-Fuzzy machine in order to model a rule-based MIMO system as logical combination of spatially distributed Single-Input Single-Output sub-systems. Each SISO systems as sensory and processing layer of the inference system, construct a single rule and learning process is handled by a Hebbian-like Spike-Time Dependent Plasticity. To shape a concrete knowledge about the whole system, extracted features of SISO neural systems (or equivalently the rules associated with SISO systems) are combined. To exhibit the system applicability, a single link cart-pole balancer as a sensory-motor learning task, has been simulated. The system is provided by reinforcement feedback from environment and is able to learn how to get expert and achieve a successful policy to perform motor control.

Mohsen Firouzi, Saeed Bagheri Shouraki, Jörg Conradt
Model-Based Identification of EEG Markers for Learning Opportunities in an Associative Learning Task with Delayed Feedback

This paper combines a reinforcement learning (RL) model and EEG data analysis to identify learning situations in a associative learning task with delayed feedback. We investigated neural correlates in occipital alpha and prefrontal theta band power of learning opportunities, identified by the RL model. We show that those parameters can also be used to differentiate between learning opportunities which lead to correct learning and those which do not. Finally, we show that learning situations can also be identified on a single trial basis.

Felix Putze, Daniel V. Holt, Tanja Schultz, Joachim Funke

Vision

Detection and Recognition

Structured Prediction for Object Detection in Deep Neural Networks

Deep convolutional neural networks are currently applied to computer vision tasks, especially object detection. Due to the large dimensionality of the output space, four dimensions per bounding box of an object, classification techniques do not apply easily. We propose to adapt a structured loss function for neural network training which directly maximizes overlap of the prediction with ground truth bounding boxes. We show how this structured loss can be implemented efficiently, and demonstrate bounding box prediction on two of the Pascal VOC 2007 classes.

Hannes Schulz, Sven Behnke
A Multichannel Convolutional Neural Network for Hand Posture Recognition

Natural communication between humans involves hand gestures, which has an impact on research in human-robot interaction. In a real-world scenario, understanding human gestures by a robot is hard due to several challenges like hand segmentation. To recognize hand postures this paper proposes a novel convolutional implementation. The model is able to recognize hand postures recorded by a robot camera in real-time, in a real-world application scenario. The proposed model was also evaluated with a benchmark database and showed better results than the ones reported in the benchmark paper.

Pablo Barros, Sven Magg, Cornelius Weber, Stefan Wermter
A Two-Stage Classifier Architecture for Detecting Objects under Real-World Occlusion Patterns

Despite extensive efforts, state-of-the-art detection approaches show a strong degradation of performance with increasing level of occlusion. In this paper we investigate a strategy to improve the detection of occluded objects based on the analytic feature framework from [11] and compare the results in a car detection task. Motivated by an analysis of annotated traffic scenes we focus on a general concept to handle vertical occlusion patterns. For this we describe a two stage classifier architecture that detects vertical car parts in the first stage and combines the local responses in the second. As an extension we provide depth information for the individual car parts helping the classifier in the second stage to reason about typical occlusion patterns.

Marvin Struwe, Stephan Hasler, Ute Bauer-Wersing
Towards Sparsity and Selectivity: Bayesian Learning of Restricted Boltzmann Machine for Early Visual Features

This paper exploits how Bayesian learning of restricted Boltzmann machine (RBM) can discover more biologically-resembled early visual features. The study is mainly motivated by the sparsity and selectivity of visual neurons’ activations in V1 area. Most previous work of computational modeling emphasize selectivity and sparsity independently, which neglects the underlying connections between them. In this paper, a prior on parameters is defined to simultaneously enhance these two properties, and a Bayesian learning framework of RBM is introduced to infer the maximum posterior of the parameters. The proposed prior performs as the lateral inhibition between neurons. According to our empirical results, the visual features learned from the proposed Bayesian framework yield better discriminative and generalization capability than the ones learned with maximum likelihood, or other state-of-the-art training strategies.

Hanchen Xiong, Sandor Szedmak, Antonio Rodríguez-Sánchez, Justus Piater

Invariances and Shape Recovery

Online Learning of Invariant Object Recognition in a Hierarchical Neural Network

We propose the

Temporal Correlation Net (TCN)

as an object recognition system implementing three basic principles: forming temporal groups of features, learning in a hierarchical structure, and using feedback to predict future input. It is a further development of the Temporal Correlation Graph [1] and shows improved performance on standard datasets like ETH80, COIL100, and ALOI. In contrast to its predecessor it can be trained online on all levels rather than in a level per level batch mode. Training images are presented in temporal order showing objects undergoing specific transformations under viewing conditions the system is supposed to learn invariance under. Computation time and memory demands are low because of sparse learned connectivity and efficient handling of neural activities.

Markus Leßmann, Rolf P. Würtz
Incorporating Scale Invariance into the Cellular Associative Neural Network

This paper describes an improvement to the Cellular Associative Neural Network, an architecture based on the distributed model of a cellular automaton, allowing it to perform scale invariant pattern matching. The use of tensor products and superposition of patterns allows the system to recall patterns at multiple resolutions simultaneously. Our experimental results show that the architecture is capable of scale invariant pattern matching, but that further investigation is needed to reduce the distortion introduced by image scaling.

Nathan Burles, Simon O’Keefe, James Austin
Shape from Shading by Model Inclusive Learning with Simultaneously Estimating Reflection Parameters

Recovering shape from shading is an important problem in computer vision and robotics and many studies have been done. We have already proposed a versatile method of solving the problem by model inclusive learning of neural networks. The method is versatile in the sense that it can solve the problem in various circumstances. Almost all of the methods of recovering shape from shading proposed so far assume that surface reflection properties of a target object are known a priori. It is, however, very difficult to obtain those properties exactly. In this paper we propose a method to resolve this problem by extending our previous method. The proposed method makes it possible to recover shape with simultaneously estimating reflection parameters of an object.

Yasuaki Kuroe, Hajimu Kawakami

Attention and Pose Estimation

Instance-Based Object Recognition with Simultaneous Pose Estimation Using Keypoint Maps and Neural Dynamics

We present a method for biologically-inspired object recognition with one-shot learning of object appearance. We use a computationally efficient model of V1 keypoints to select object parts with the highest information content and model their surroundings using simple colour features. This map-like representation is fed into a dynamical neural network which performs pose, scale and translation estimation of the object given a set of previously observed object views. We demonstrate the feasibility of our algorithm for cognitive robotic scenarios and evaluate classification performance on a dataset of household items.

Oliver Lomp, Kasim Terzić, Christian Faubel, J. M. H. du Buf, Gregor Schöner
How Visual Attention and Suppression Facilitate Object Recognition?

Visual attention can support object recognition by selecting the relevant target information in the huge amount of sensory data, especially important in scenes composed of multiple objects. Here we demonstrate how attention in a biologically plausible and neuro-computational model of visual perception facilitates object recognition in a robotic real world scenario. We will point out that it is not only important to select the target information, but rather to explicitly suppress the distracting sensory data. We found that suppressing the features of each distractor is not sufficient to achieve robust recognition. Instead, we also have to suppress the location of each distractor. To demonstrate the effect of this spatial suppression, we disable this property and show that the recognition accuracy drops. By this, we show the interplay between attention and suppression in a real world object recognition task.

Frederik Beuth, Amirhossein Jamalian, Fred H. Hamker
Analysis of Neural Circuit for Visual Attention Using Lognormally Distributed Input

Visual attention has recently been reported to modulate neural activity of narrow spiking and broad spiking neurons in V4, with increased firing rate and less inter-trial variations. We simulated these physiological phenomena using a neural network model based on spontaneous activity, assuming that the visual attention modulation could be achieved by a change in variance of input firing rate distributed with a lognormal distribution. Consistent with the physiological studies, an increase in firing rate and a decrease in inter-trial variance was simultaneously obtained in the simulation by increasing variance of input firing rate distribution. These results indicate that visual attention forms strong sparse and weak dense input or a ‘winner-take-all’ state, to improve the signal-to-noise ratio of the target information.

Yoshihiro Nagano, Norifumi Watanabe, Atsushi Aoyama

Supervised Learning

Ensembles

Dynamic Ensemble Selection and Instantaneous Pruning for Regression Used in Signal Calibration

A dynamic method of selecting a pruned ensemble of predictors for regression problems is described. The proposed method enhances the prediction accuracy and generalization ability of pruning methods that change the order in which ensemble members are combined. Ordering heuristics attempt to combine accurate yet complementary regressors. The proposed method enhances the performance by modifying the order of aggregation through distributing the regressor selection over the entire dataset. This paper compares four static ensemble pruning approaches with the proposed dynamic method. The experimental comparison is made using MLP regressors on benchmark datasets and on an industrial application of radio frequency source calibration.

Kaushala Dias, Terry Windeatt
Global and Local Rejection Option in Multi–classification Task

This work presents two rejection options. The global rejection option separates the foreign observations – not defined in the classification task – from the normal observations. The local rejection option works after the classification process and separates observations individually for each class. We present implementation of both methods for binary classifiers grouped in a graph structure (tree or directed acyclic graph). Next, we prove that the quality of rejection is identical for both options and depends only on the quality of binary classifiers. The methods are compared on the handwritten digits recognition task. The local rejection option works better for the most part.

Marcin Luckner
Comparative Study of Accuracies on the Family of the Recursive-Rule Extraction Algorithm

In this paper, we first compare the accuracies of the Recursive-Rule Extraction algorithm family, i.e., the Re-RX algorithm, its variant and the “Three-MLP Ensemble by the Re-RX algorithm” (shortened to “Three-MLP Ensemble”) using the Re-RX algorithm as a core part for six kinds of two-class mixed (i.e., discrete and continuous attributes) datasets. Two-class mixed datasets are commonly used for credit scoring and generally in financial domains. In this paper, we compare the accuracy by only the Re-RX algorithm family because of recent comparison reviews and benchmarking study results, obtained by complicated statistics, support vector machines, neuro-fuzzy hybrid classifications, and similar techniques. The Three-MLP Ensemble algorithm cascades standard backpropagation (BP) to train a three neural-network ensemble, where each neural network is a Multi-Layer Perceptron (MLP). Thus, strictly speaking, three neural networks do not need to be trained simultaneously. In addition, the Three-MLP Ensemble is a simple and new concept of rule extraction from neural network ensembles and can avoid previous complicated neural network ensemble structures and the difficulties of rule extraction algorithms. The extremely high accuracy of the Three-MLP Ensemble algorithm generally outperformed the Re-RX algorithm and the variant. The results confirm that the output from the network ensemble can be expressed in the form of rules, and thus opens the “black box” of trained neural network ensembles.

Yoichi Hayashi, Yuki Tanaka, Shota Fujisawa, Tomoki Izawa
Improving the Convergence Property of Soft Committee Machines by Replacing Derivative with Truncated Gaussian Function

In online gradient descent learning, the local property of the derivative of the output function can cause slow convergence. This phenomenon, called a

plateau

, occurs in the learning process of a multilayer network. Improving the derivative term, we propose a simple method replacing the derivative term with a truncated Gaussian function that greatly increases the convergence speed. We then analyze a soft committee machine trained by proposed method, and show how proposed method breaks a plateau. Results showed that the proposed method eventually led to break the symmetry between hidden units.

Kazuyuki Hara, Kentaro Katahira

Regression

Fast Sensitivity-Based Training of BP-Networks

Sensitivity analysis became an acknowledged tool used to study the performance of artificial neural networks. Sensitivity analysis allows to assess the influence, e.g., of each neuron or weight on the final network output. In particular various feature selection and pruning strategies are based on this capability. In this paper, we will present a new approximative sensitivity-based training algorithm yielding robust neural networks with generalization capabilities comparable to its exact analytical counterpart, yet much faster.

Iveta Mrázová, Zuzana Petříčková
Learning Anisotropic RBF Kernels

We present an approach for learning an anisotropic RBF kernel in a game theoretical setting where the value of the game is the degree of separation between positive and negative training examples. The method extends a previously proposed method (KOMD) to perform feature re-weighting and distance metric learning in a kernel-based classification setting. Experiments on several benchmark datasets demonstrate that our method generally outperforms state-of-the-art distance metric learning methods, including the Large Margin Nearest Neighbor Classification family of methods.

Fabio Aiolli, Michele Donini
Empowering Imbalanced Data in Supervised Learning: A Semi-supervised Learning Approach

We present a framework to address the imbalanced data problem using semi-supervised learning. Specifically, from a supervised problem, we create a semi-supervised problem and then use a semi-supervised learning method to identify the most relevant instances to establish a well-defined training set. We present extensive experimental results, which demonstrate that the proposed framework significantly outperforms all other sampling algorithms in 67% of the cases across three different classifiers and ranks second best for the remaining 33% of the cases.

Bassam A. Almogahed, Ioannis A. Kakadiaris
A Geometrical Approach for Parameter Selection of Radial Basis Functions Networks

The RBF network is commonly used for classification and function approximation. The center and radius of the activation function of neurons is an important parameter to be found before the network training. This paper presents a method based on computational geometry to find these coefficients without any parameters provided by the user. The method is compared with a SVM and experimental results showed that our approach is promising.

Luiz C. B. Torres, André P. Lemos, Cristiano L. Castro, Antônio P. Braga
Sampling Hidden Parameters from Oracle Distribution

A new sampling learning method for neural networks is proposed. Derived from an integral representation of neural networks, an

oracle

probability distribution of hidden parameters is introduced. In general rigorous sampling from the oracle distribution holds numerical difficulty, a linear-time sampling algorithm is also developed. Numerical experiments showed that when hidden parameters were initialized by the oracle distribution, following backpropagation converged faster to better parameters than when parameters were initialized by a normal distribution.

Sho Sonoda, Noboru Murata
Incremental Input Variable Selection by Block Addition and Block Deletion

In selecting input variables by block addition and block deletion (BABD), multiple input variables are added and then deleted, keeping the cross-validation error below that using all the input variables. The major problem of this method is that selection time becomes large as the number of input variables increases. To alleviate this problem, in this paper, we propose incremental block addition and block deletion of input variables. In this method, for an initial subset of input variables we select input variables by BABD. Then in the incremental step, we add some input variables that are not added before to the current selected input variables and iterate BABD. To guarantee that the cross-validation error decreases monotonically by incremental BABD, we undo incremental BABD if the obtained cross-validation error rate is worse than that at the previous incremental step. We evaluate incremental BABD using some benchmark data sets and show that by incremental BABD, input variable selection is speeded up with the approximation error comparable to that by batch BABD.

Shigeo Abe

Classification

A CFS-Based Feature Weighting Approach to Naive Bayes Text Classifiers

Recent work in supervised learning has shown that naive Bayes text classifiers with strong assumptions of independence among features, such as multinomial naive Bayes (MNB), complement naive Bayes (CNB) and the one-versus-all-but-one model (OVA), have achieved remarkable classification performance. This fact raises the question of whether a naive Bayes text classifier with less restrictive assumptions can perform even better. Responding to this question, we firstly evaluate the correlation-based feature selection (CFS) approach in this paper and find that it performs even worse than the original versions. Then, we propose a CFS-based feature weighting approach to these naive Bayes text classifiers. We call our feature weighted versions FWMNB, FWCNB and FWOVA respectively. Our proposed approach weakens the strong assumptions of independence among features by weighting the correlated features. The experimental results on a large suite of benchmark datasets show that our feature weighted versions significantly outperform the original versions in terms of classification accuracy.

Shasha Wang, Liangxiao Jiang, Chaoqun Li
Local Rejection Strategies for Learning Vector Quantization

Classification with rejection is well understood for classifiers which provide explicit class probabilities. The situation is more complicated for popular deterministic classifiers such as learning vector quantisation schemes: albeit reject options using simple distance-based geometric measures were proposed [4], their local scaling behaviour is unclear for complex problems. Here, we propose a local threshold selection strategy which automatically adjusts suitable threshold values for reject options in prototype-based classifiers from given data. We compare this local threshold strategy to a global choice on artificial and benchmark data sets; we show that local thresholds enhance the classification results in comparison to global ones, and they better approximate optimal Bayesian rejection in cases where the latter is available.

Lydia Fischer, Barbara Hammer, Heiko Wersing
Efficient Adaptation of Structure Metrics in Prototype-Based Classification

More complex data formats and dedicated structure metrics have spurred the development of intuitive machine learning techniques which directly deal with dissimilarity data, such as relational learning vector quantization (RLVQ). The adjustment of metric parameters like relevance weights for basic structural elements constitutes a crucial issue therein, and first methods to automatically learn metric parameters from given data were proposed recently. In this contribution, we investigate a robust learning scheme to adapt metric parameters such as the scoring matrix in sequence alignment in conjunction with prototype learning, and we investigate the suitability of efficient approximations thereof.

Bassam Mokbel, Benjamin Paassen, Barbara Hammer
Improved Adaline Networks for Robust Pattern Classification

The Adaline network [1] is a classic neural architecture whose learning rule is the famous least mean squares (LMS) algorithm (a.k.a. delta rule or Widrow-Hoff rule). It has been demonstrated that the LMS algorithm is optimal in

H

 ∞ 

sense since it tolerates

small

(in energy) disturbances, such as measurement noise, parameter drifting and modelling errors [2,3]. Such optimality of the LMS algorithm, however, has been demonstrated for regression-like problems only, not for pattern classification. Bearing this in mind, we firstly show that the performances of the LMS algorithm and variants of it (including the recent Kernel LMS algorithm) in pattern classification tasks deteriorates considerably in the presence of labelling errors, and then introduce robust extensions of the Adaline network that can deal efficiently with such errors. Comprehensive computer simulations show that the proposed extension consistently outperforms the original version.

César Lincoln C. Mattos, José Daniel A. Santos, Guilherme A. Barreto
Learning under Concept Drift with Support Vector Machines

Support Vector Machines (SVMs) have been recognized as one of the most successful classification methods for many applications in static environment. However in dynamic environment, data characteristics may evolve over time. This leads to deteriorate dramatically the performance of SVMs over time. This is because of the use of data which is no more consistent with the characteristics of new incoming one. Thus in this paper, we propose an approach to recognize and handle concept changes with support vector machine. This approach integrates a mechanism to use only the recent and most representative patterns to update the SVMs without a catastrophic forgetting.

Omar Ayad
Two Subspace-Based Kernel Local Discriminant Embedding

We propose Two Subspace-based Kernel Local Discriminant Embedding (TSKLDE) for feature extraction and recognition. The procedure of TSKLDE is divided into two stages. The first stage applies the Kernel Principal Component Analysis (KPCA) to the raw data. Based on the output of KPCA, the second stage seeks two kinds of discriminant information, regular and irregular. These are based on transforms derived from the within-class locality preserving scatter. The resulting framework retain the advantages of supervised global and local techniques. Besides, it is an inductive nonlinear technique in the sense that novel examples or images can be directly mapped. The proposed algorithm was tested and evaluated using three public face databases. The experimental results show that the proposed TSKLDE outperforms other Kernel based algorithms.

Fadi Dornaika, Alireza Bosaghzadeh

Dynamical Models and Time Series

Coupling Gaussian Process Dynamical Models with Product-of-Experts Kernels

We describe a new probabilistic model for learning of coupled dynamical systems in latent state spaces. The coupling is achieved by combining predictions from several Gaussian process dynamical models in a product-of-experts fashion. Our approach facilitates modulation of coupling strengths without the need for computationally expensive re-learning of the dynamical models. We demonstrate the effectiveness of the new coupling model on synthetic toy examples and on high-dimensional human walking motion capture data.

Dmytro Velychko, Dominik Endres, Nick Taubert, Martin A. Giese
A Deep Dynamic Binary Neural Network and Its Application to Matrix Converters

This paper studies the deep dynamic binary neural network that is characterized by the signum activation function, ternary weighting parameters and integer threshold parameters. In order to store a desired binary periodic orbit, we present a simple learning method based on the correlation learning. The method is applied to a teacher signal that corresponds to control signal of the matrix converter in power electronics. Performing numerical experiments, we investigate storage of the teacher signal and its stability as the depth of the network varies.

Jungo Moriyasu, Toshimichi Saito
Improving Humanoid Robot Speech Recognition with Sound Source Localisation

In this paper we propose an embodied approach to automatic speech recognition, where a humanoid robot adjusts its orientation to the angle that increases the signal-to-noise ratio of speech. In other words, the robot turns its face to ’hear’ the speaker better, similar to what people with auditory deficiencies do. The robot tracks a speaker with a binaural sound source localisation system (SSL) that uses spiking neural networks to model relevant areas in the mammalian auditory pathway for SSL. The accuracy of speech recognition is doubled when the robot orients towards the speaker in an optimal angle and listens only through one ear instead of averaging the input from both ears.

Jorge Dávila-Chacón, Johannes Twiefel, Jindong Liu, Stefan Wermter
Control of UPOs of Unknown Chaotic Systems via ANN

In this paper an artificial neural network (ANN) is developed for modeling and controlling unknown chaotic systems to unstable periodic orbits (UPOs). In the modeling phase, the ANN is trained on the unknown chaotic systems using the input-output data obtained from the unknown (or uncertain) underlying chaotic systems, and a specific computational algorithm is employed for the parameter optimization. In the controlling phase, the

L

2

-stability criterion is used, which forms the basis of the main design principle. Some simulation results on the chaotic Henon and Duffing systems are given, for both modeling and controlling phases, to illustrate the effectiveness of the proposed chaos control scheme and the proposed neural network.

Amine M. Khelifa, Abdelkrim Boukabou
Event-Based Visual Data Sets for Prediction Tasks in Spiking Neural Networks

For spiking networks to perform computational tasks, benchmark data sets are required for model design, refinement and testing. Classic machine learning benchmark data sets use classification as the dominant paradigm, however the temporal characteristics of spiking neural networks mean they are likely to be more useful for problems involving sequence data. To support these paradigms, we provide data sets of 11 moving scenes, each with multiple variations, recorded from a dynamic vision sensor (DVS128), comprising high dimensional (16k pixels) and low latency (15 microsecond) events. We also present a novel long range prediction task based on the DVS128 data, and introduce a pilot study of a spiking neural network learning to predict thousands of events into the future.

Tingting (Amy) Gibson, Scott Heath, Robert P. Quinn, Alexia H. Lee, Joshua T. Arnold, Tharun S. Sonti, Andrew Whalley, George P. Shannon, Brian T. Song, James A. Henderson, Janet Wiles
Modeling of Chaotic Time Series by Interval Type-2 NEO-Fuzzy Neural Network

This paper describes the development of Interval Type-2 NEO-Fuzzy Neural Network for modeling of complex dynamics. The proposed network represents a parallel set of multiple zero order Sugeno type approximations, related only to their own input argument. The induced gradient based learning procedure, adjusts solely the consequent network parameters. To improve the robustness of the network and the possibilities for handling uncertainties, Type-2 Gaussian fuzzy sets are introduced into the network topology. The potentials of the proposed approach in modeling of Mackey-Glass and Rossler Chaotic time series are studied.

Yancho Todorov, Margarita Terziyska

Neuroscience

Cortical Models

Excitation/Inhibition Patterns in a System of Coupled Cortical Columns

We study how excitation and inhibition are distributed mesoscopically in small brain regions, by means of a computational model of coupled cortical columns described by neural mass models. Two cortical columns coupled bidirectionally through both excitatory and inhibitory connections can spontaneously organize in a regime in which one of the columns is purely excitatory and the other is purely inhibitory, provided the excitatory and inhibitory coupling strengths are adequately tuned. We also study the case of three columns in different coupling configurations (linear array and all-to-all coupling), finding abrupt transitions between heterogeneous and homogeneous excitatory/inhibitory patterns and strong multistability in their distribution.

Daniel Malagarriga, Alessandro E. P. Villa, Jordi García-Ojalvo, Antonio J. Pons
Self-generated Off-line Memory Reprocessing Strongly Improves Generalization in a Hierarchical Recurrent Neural Network

Strong experimental evidence suggests that cortical memory traces are consolidated during off-line memory reprocessing that occurs in the off-line states of sleep or waking rest. It is unclear, what plasticity mechanisms are involved in this process and what changes are induced in the network in the off-line regime. Here, we examine a hierarchical recurrent neural network that performs unsupervised learning on natural face images of different persons. The proposed network is able to self-generate memory replay while it is decoupled from external stimuli. Remarkably, the recognition performance is tremendously boosted after this off-line regime specifically for the novel face views that were not shown during the initial learning. This effect is independent of synapse-specific plasticity, relying completely on homeostatic regulation of intrinsic excitability. Comparing a purely feed-forward network configuration with the full version reveals a substantially stronger boost in recognition performance for the fully recurrent network architecture after the off-line regime.

Jenia Jitsev
Lateral Inhibition Pyramidal Neural Networks Designed by Particle Swarm Optimization

LIPNet is a pyramidal neural network with lateral inhibition developed for pattern recognition, inspired in the concept of receptive and inhibitory fields from the human visual system. Although this network can implicitly extract features and use these features to properly classify patterns in images, many parameters must be defined prior to the network training and operation. Besides, these parameters have a huge impact on the recognition performance. This paper proposes an encoding scheme aiming at optimizing the LIPNet structure using Particle Swarm Optimization. Preliminary results for a face detection problem using a well known benchmark set showed that our approach achieved better classification rates when compared to the original LIPNet.

Alessandra M. Soares, Bruno J. T. Fernandes, Carmelo J. A. Bastos-Filho
Bio-mimetic Path Integration Using a Self Organizing Population of Grid Cells

Grid cells in the dorsocaudal medial entorhinal cortex (dMEC) of the rat provide a metric representation of the animal’s local environment. The collective firing patterns in a network of grid cells forms a triangular mesh that accurately tracks the location of the animal. The activity of a grid cell network, similar to head direction cells, displays path integration characteristics. Classical robotics use path integrators in the form of inertial navigation systems to track spatial information of an agent as well. In this paper, we describe an implementation of a network of grid cells as a dead reckoning system for the PR2 robot.

Ankur Sinha, Jack Jianguo Wang
Learning Spatial Transformations Using Structured Gain-Field Networks

Brains experience sensory information grounded in sensor- relative frames of reference. To compare sensory information from different sensor sources, such as vision and touch, this information needs to be mapped onto each other. To do so, the brain needs to learn suitable spatial transformations and the literature suggests that gain fields accomplish such transformations. However, when transforming three dimensional spaces or even six dimensional configuration spaces then simple gain fields do not scale to such a dimensionality. We are investigating how this curse of dimensionality can be overcome. Based on neural population-encoded, component-wise spatial representations, we show that a hierarchy of gain fields can accomplish higher-dimensional transformations and that its weights can be learned effectively by means of standard backpropagation.

Jan Kneissler, Martin V. Butz

Line Attractors and Neural Fields

Flexible Cue Integration by Line Attraction Dynamics and Divisive Normalization

One of the key computations performed in human brain is multi-sensory cue integration, through which humans are able to estimate the current state of the world to discover relative reliabilities and relations between observed cues. Mammalian cortex consists of highly distributed and interconnected populations of neurons, each providing a specific type of information about the state of the world. Connections between areas seemingly realize functional relationships amongst them and computation occurs by each area trying to be consistent with the areas it is connected to. In this paper using line-attraction dynamics and divisive normalization, we present a computational framework which is able to learn arbitrary non-linear relations between multiple cues using a simple Hebbian Learning principle. After learning, the network dynamics converges to the stable state so to satisfy the relation between connected populations. This network can perform several principle computational tasks such as inference, de-noising and cue-integration. By applying a real world multi-sensory integrating scenario, we demonstrate that the network can encode relative reliabilities of cues in different areas of the state space, over distributed population vectors. This reliability based encoding biases the network’s dynamics in favor of more reliable cues and realizes a near optimal sensory integration mechanism. Additional important features of the network are its scalability to cases with higher order of modalities and its flexibility to learn smooth functions of relations which is necessary for a system to operate in a dynamic environment.

Mohsen Firouzi, Stefan Glasauer, Jörg Conradt
Learning to Look: A Dynamic Neural Fields Architecture for Gaze Shift Generation

Looking is one of the most basic and fundamental goal-directed behaviors. The neural circuitry that generates gaze shifts towards target objects is adaptive and compensates for changes in the sensorimotor plant. Here, we present a neural-dynamic architecture, which enables an embodied agent to direct its gaze towards salient objects in its environment. The sensorimotor mapping, which is needed to accurately plan the gaze shifts, is initially learned and is constantly updated by a gain adaptation mechanism. We implemented the architecture in a simulated robotic agent and demonstrated autonomous map learning and adaptation in an embodied setting.

Christian Bell, Tobias Storck, Yulia Sandamirskaya
Skeleton Model for the Neurodynamics of Visual Action Representations

The visual recognition of body motion in the primate brain requires the temporal integration of information over complex patterns, potentially exploiting recurrent neural networks consisting of shape- and optic-flow-selective neurons. The paper presents a mathematically simple neurodynamical model that approximates the mean-field dynamics of such networks. It is based on a two-dimensional neural field with appropriate lateral interaction kernel and an adaptation process for the individual neurons. The model accounts for a number of, so far not modeled, observations in the recognition of body motion, including perceptual multi-stability and the weakness of repetition suppression, as observed in single-cell recordings for the repeated presentation of action stimuli. In addition, the model predicts novel effects in the perceptual organization of action stimuli.

Martin A. Giese
Latency-Based Probabilistic Information Processing in Recurrent Neural Hierarchies

In this article, we present an original neural space/latency code, integrated in a multi-layered neural hierarchy, that offers a new perspective on probabilistic inference operations. Our work is based on the dynamic neural field paradigm that leads to the emergence of activity bumps, based on recurrent lateral interactions, thus providing a spatial coding of information. We propose that lateral connections represent a data model, i.e., the conditional probability of a “true” stimulus given a noisy input. We propose furthermore that the resulting attractor state encodes the most likely ”true” stimulus given the data model, and that its latency expresses the confidence in this interpretation. Thus, the main feature of this network is its ability to represent, transmit and integrate probabilistic information at multiple levels so that to take near-optimal decisions when inputs are contradictory, noisy or missing. We illustrate these properties on a three-layered neural hierarchy receiving inputs from a simplified robotic object recognition task. We also compare the network dynamics to an explicit probabilistic model of the task, to verify that it indeed reproduces all relevant properties of probabilistic processing.

Alexander Gepperth, Mathieu Lefort

Spiking and Single Cell Models

Factors Influencing Polychronous Group Sustainability as a Model of Working Memory

Several computational models have been designed to help our understanding of the conditions under which persistent activity can be sustained in cortical circuits during working memory tasks. Here we focus on one such model that has shown promise, that uses polychronization and short term synaptic dynamics to achieve this reverberation, and explore it with respect to different physiological parameters in the brain, including size of the network, number of synaptic connections, small-world connectivity, maximum axonal conduction delays, and type of cells (excitatory or inhibitory). We show that excitation and axonal conduction delays greatly affect the sustainability of spatio-temporal patterns of spikes called polychronous groups.

Panagiotis Ioannou, Matthew Casey, André Grüning
Pre- and Postsynaptic Properties Regulate Synaptic Competition through Spike-Timing-Dependent Plasticity

Brain functions such as learning and memory rely on synaptic plasticity. Many studies have shown that synaptic plasticity can be driven by the timings between pre- and postsynaptic spikes, also known as spike-timing-dependent plasticity (STDP). In most of the modeling studies exploring STDP functions, presynaptic spikes have been postulated to be Poisson (random) spikes and postsynaptic neurons have been described using an integrate-and-fire model, for simplicity. However, experimental data suggest this is not necessarily true. In this study, we investigated how STDP worked in synaptic competition if more neurophysiologically realistic properties for pre- and postsynaptic dynamics were incorporated; presynaptic (input) spikes obeyed a gamma process and the postsynaptic neuron was a multi-timescale adaptive threshold model. Our results showed that STDP strengthened specific combinations of pre- and postsynaptic properties; regular spiking neurons favored regular input spikes whereas random spiking neurons did random ones, suggesting neural information coding utilizes both the properties.

Hana Ito, Katsunori Kitano
Location-Dependent Dendritic Computation in a Modeled Striatal Projection Neuron

The striatum comprises part of a feedback loop between the cerebral cortex, thalamus and other nuclei of the basal ganglia, ultimately guiding action selection and motor learning. Much of this is facilitated by striatal projection neurons, which receive and process highly convergent cortical and thalamic excitatory inputs. All of the glutamatergic inputs to projection neurons synapse on dendrites, many directly on spine heads. The distal, but not proximal, dendrites of projection neurons are capable of supporting synaptically driven regenerative events, which are transfered to the soma as depolarized upstates from which action potentials can occur. In this study we present a modified NEURON model of a striatal projection neuron, and use it to examine the location-dependence of upstate generation and action potential gating. Specifically, simulations show that the small diameter of distal SPN dendrites can support plateau potentials by increasing the cooperativity among neighboring spines. Furthermore, such distally evoked plateaus can boost the somatic response to stimulation of proximal dendritic spines, facilitating action potential generation. The implications these results have for action selection are discussed.

Youwei Zheng, Lars Schwabe, Joshua L. Plotkin
Classifying Spike Patterns by Reward-Modulated STDP

Reward-modulated learning rules for spiking neural networks have emerged, that have been demonstrated to solve a wide range of reinforcement learning tasks. Despite this, little work has aimed to classify spike patterns by the timing of output spikes. Here, we apply a rewardmaximising learning rule to teach a spiking neural network to classify input patterns by the latency of output spikes. Furthermore, we compare the performance of two escape rate functions that drive output spiking activity: the Arrhenius & Current (A&C) model and Exponential (EXP) model. We find A&C consistently outperforms EXP, and especially in terms of the time taken to converge in learning. We also show that jittering input patterns with a low noise amplitude leads to an improvement in learning, by reducing the variation in the performance.

Brian Gardner, Ioana Sporea, André Grüning

Applications

Users and Social Technologies

Quantifying the Effect of Meaning Variation in Survey Analysis

Surveys are widely conducted as a means to obtain information on thoughts, opinions and feelings of people. The representativeness of a sample is a major concern in using surveys. In this article, we consider meaning variation which is another potentially remarkable but less studied source of problems. We use Grounded Intersubjective Concept Analysis (GICA) method to quantify meaning variation and demonstrate the effect on survey analysis through a case study in which food prices and food concepts are considered.

Henri Sintonen, Juha Raitio, Timo Honkela
Discovery of Spatio-Temporal Patterns from Foursquare by Diffusion-type Estimation and ICA

In this paper, we extract various patterns of the spatio-temporal distribution from Foursquare. Foursquare is a location-based social networking system which has been widely used recently. For extracting patterns, we employ ICA (Independent Component Analysis), which is a useful method in signal processing and feature extraction. Because the Foursquare dataset consists of check-in’s of users at some time points and locations, ICA is not directly applicable to it. In order to smooth the dataset, we estimate a continuous spatio-temporal distribution by employing a diffusion-type formula. The experiments on an actual Foursquare dataset showed that the proposed method could extract some plausible and interesting spatio-temporal patterns.

Yoshitatsu Matsuda, Kazunori Yamaguchi, Ken-ichiro Nishioka
Content-Boosted Restricted Boltzmann Machine for Recommendation

Collaborative filtering and Content-based filtering methods are two famous methods used by recommender systems. Restricted Boltzmann Machine(RBM) model rivals the best collaborative filtering methods, but it focuses on modeling the correlation between item ratings. In this paper, we extend RBM model by incorporating content-based features such as user demograohic information, items categorization and other features. We use Naive Bayes classifier to approximate the missing entries in the user-item rating matrix, and then apply the modified UI-RBM on the denser rating matrix. We present expermental results that show how our approach, Content-boosted Restricted Boltzmann Machine(CB-RBM), performs better than a pure RBM model and other content-boosted collaborative filtering methods.

Yongqi Liu, Qiuli Tong, Zhao Du, Lantao Hu
Financial Self-Organizing Maps

This paper introduces Financial Self–Organizing Maps (FinSOM) as a SOM sub–class where the mapping of inputs on the neural space takes place using functions with economic soundness, that makes them particularly well–suited to analyze financial data. The visualization capabilities as well as the explicative power of both the standard SOM and the FinSOM variants is tested on data from the German Stock Exchange. The results suggest that, dealing with financial data, the FinSOM seem to offer superior representation capabilities of the observed phenomena.

Marina Resta

Technical Systems

RatSLAM on Humanoids - A Bio-Inspired SLAM Model Adapted to a Humanoid Robot

Mapping, localization and navigation are major topics and challenges for mobile robotics. To perform tasks and to interact efficiently in the environment, a robot needs knowledge about its surroundings. Many robots today are capable of performing simultaneous mapping and localization to generate own world representations. Most assume an array of highly sophisticated artificial sensors to track landmarks placed in the environment. Recently, there has been significant interest in research approaches inspired by nature and RatSLAM is one of them. It has been introduced and tested on wheeled robots with good results. To examine how RatSLAM behaves on humanoid robots, we adapt this model for the first time to this platform by adjusting the given constraints. Furthermore, we introduce a multiple hypotheses mapping technique which improves mapping robustness in open spaces with features visible from several distant locations.

Stefan Müller, Cornelius Weber, Stefan Wermter
Precise Wind Power Prediction with SVM Ensemble Regression

In this work, we propose the use of support vector regression ensembles for wind power prediction. Ensemble methods often yield better classification and regression accuracy than classical machine learning algorithms and reduce the computational cost. In the field of wind power generation, the integration into the smart grid is only possible with a precise forecast computed in a reasonable time. Our support vector regression ensemble approach uses bootstrap aggregating (bagging), which can easily be parallelized. A set of weak predictors is trained and then combined to an ensemble by aggregating the predictions. We investigate how to choose and train the individual predictors and how to weight them best in the prediction ensemble. In a comprehensive experimental analysis, we show that our SVR ensemble approach renders significantly better forecast results than state-of-the-art predictors.

Justin Heinermann, Oliver Kramer
Neural Network Approaches to Solution of the Inverse Problem of Identification and Determination of Partial Concentrations of Salts in Multi-сomponent Water Solutions

The studied inverse problem is determination of partial concentrations of inorganic salts in multi-component water solutions by their Raman spectra. The problem is naturally divided into two parts: 1) determination of the component composition of the solution, i.e. which salts are present and which not; 2) determination of the partial concentration of each of the salts present in the solution. Within the first approach, both parts of the problem are solved simultaneously, with a single neural network (perceptron) with several outputs, each of them estimating the concentration of the corresponding salt. The second approach uses data clusterization by Kohonen networks for consequent identification of component composition of the solution by the cluster, which the spectrum of this solution falls into. Both approaches and their results are discussed in this paper.

Sergey Dolenko, Sergey Burikov, Tatiana Dolenko, Alexander Efitorov, Kirill Gushchin, Igor Persiantsev
Lateral Inhibition Pyramidal Neural Network for Detection of Optical Defocus (Zernike Z 5)

Optical distortions of an image are created in astronomical, optical microscopy and communication systems by light propagation throughout a variety of optical components. Usually, optical aberrations are corrected by using an Adaptive Optics system, where a wavefront sensor is used to measure the optical distortion. In this work, we propose to use a Lateral Inhibition Pyramidal Neural Network (LIPNet) in the frequency domain to classify optical defocus (using Zernike polynomials

Z

5

), such that optical defocus value can be detected directly from the image without the use of wavefront sensing. The results show the potentiality of the method and open new opportunities to explore this kind of neural networks algorithms for wavefront sensing and Adaptive optic systems.

Bruno J. T. Fernandes, Diego Rativa
Development of a Dynamically Extendable SpiNNaker Chip Computing Module

The SpiNNaker neural computing project has created a hardware architecture capable of scaling up to a system with more than a million embedded cores, in order to simulate more than one billion spiking neurons in biological real time. The heart of this system is the SpiNNaker chip, a multi-processor System-on-Chip with a high level of interconnectivity between its processing units. Here we present a Dynamically Extendable SpiNNaker Chip Computing Module that allows a SpiNNaker machine to be deployed on small mobile robots. A non-neural application, the simulation of the movement of a flock of birds, was developed to demonstrate the general purpose capabilities of this new platform. The developed SpiNNaker machine allows the simulation of up to one million spiking neurons in real time with a single SpiNNaker chip and is scalable up to 256 computing nodes in its current state.

Rui Araújo, Nicolai Waniek, Jörg Conradt
The Importance of Physiological Noise Regression in High Temporal Resolution fMRI

Recently a new technique called multiband imaging was introduced, it allows extremely low repetition times for functional magnetic resonance imaging (fMRI). As these ultra fast imaging scans can increase the Nyquist rate by an order of magnitude, there are many new effects, that have to be accounted for. As more frequencies can now be sampled directly, we want to analyze especially those that are due to physiological noise, such as cardiac and respiratory signals. Here, we adapted RETROICOR [4] to handle multiband fMRI data. We show the importance of physiological noise regression for standard temporal resolution fMRI and compare it to the high temporal resolution case. Our results show that especially for multiband fMRI scans, it is of the utmost importance to apply physiological noise regression, as residuals of these noises are clearly detectable in non noise independent components if no prior physiological noise has been applied.

Norman Scheel, Catie Chang, Amir Madany Mamlouk
Development of Automated Diagnostic System for Skin Cancer: Performance Analysis of Neural Network Learning Algorithms for Classification

Melanoma is the most deathly of all skin cancers but early diagnosis can ensure a high degree of survival. Early diagnosis is one of the greatest challenges due to lack of experience of general practitioners (GPs). In this paper we present a clinical decision support system designed for general practitioners, aimed at saving time and resources in the diagnostic process. Segmentation, pattern recognition, and change detection are the important steps in our approach. This paper also investigates the performance of Artificial Neural Network (ANN) learning algorithms for skin cancer diagnosis. The capabilities of three learning algorithms i.e. Levenberg-Marquardt (LM), Resilient Back propagation (RP), Scaled Conjugate Gradient (SCG) algorithms in differentiating melanoma and benign lesions are studied and their performances are compared. The results show that Levenberg-Marquardt algorithm was quick and efficient in figuring out benign lesions with specificity 95.1%, while SCG algorithm gave better results in detecting melanoma at the cost of more number of epochs with sensitivity 92.6%.

Ammara Masood, Adel Ali Al-Jumaily, Tariq Adnan
Backmatter
Metadata
Title
Artificial Neural Networks and Machine Learning – ICANN 2014
Editors
Stefan Wermter
Cornelius Weber
Włodzisław Duch
Timo Honkela
Petia Koprinkova-Hristova
Sven Magg
Günther Palm
Alessandro E. P. Villa
Copyright Year
2014
Publisher
Springer International Publishing
Electronic ISBN
978-3-319-11179-7
Print ISBN
978-3-319-11178-0
DOI
https://doi.org/10.1007/978-3-319-11179-7

Premium Partner