New monograph

Andrzej Cichocki cia at brain.riken.go.jp
Fri May 31 13:01:51 EDT 2002


[Our sincere apologies if you receive multiple copies of this email]

The following book is now available:

 ADAPTIVE BLIND SIGNAL and IMAGE PROCESSING:
     Learning Algorithms and Applications
     A. Cichocki,  S. Amari

Published by John Wiley & Sons, Chichester UK,
April 2002, 586 Pages.

The books cover the  following areas:
Independent Component Analysis (ICA),  blind source separation (BSS), 
blind recovery, blind signal extraction (BSE),
multichannel blind deconvolution, blind equalization, second and higher 
order statistics, blind spatial and temporal decorrealtion, robust 
whitening,
blind filtering,  matrix factorizations,  robust principal component 
analysis, minor component analysis, sparse  representations,
automatic dimension reduction, features extraction in high dimensional 
data, noise reduction and related problems.
Moreover, some interesting  benchmarks are available to compare
performance of various unsupervised learning algorithms.

More information about the book you can find on web pages:

http://www.bsp.brain.riken.go.jp/ICAbookPAGE/
http://www.wiley.com/cda/product/0,,0471607916,00.html

and in below brief summary.


Andrzej Cichocki
Laboratory for Advanced Brain Signal Processing,
Riken BSI
2-1 Hirosawa, Wako-shi, Saitama 351-0198,
JAPAN
E-mail: cia at bsp.brain.riken.go.jp
URL: http://www.bsp.brain.riken.go.jp/




Summary of the book

Chapter 1: Introduction to Blind Signal Processing: Problems and
Applications

Blind Signal Processing (BSP) is now one of the hottest and exciting 
topics in the fields of neural computation, advanced statistics, and 
signal processing with solid theoretical foundations and many potential 
applications. In fact, BSP has become a very important topic of research 
and development in many areas, especially biomedical engineering, 
medical imaging, speech enhancement, remote sensing, communication 
systems, exploration seismology, geophysics, econometrics, data mining, 
neural networks, etc.  The blind signal processing techniques 
principally do not use any training data and do not assume a priori 
knowledge about parameters of convolutive, filtering and mixing systems. 
BSP includes three major areas: Blind Signal Separation and Extraction 
(BSS/BSE), Independent Component Analysis (ICA), and Multichannel Blind 
Deconvolution (MBD) and Equalization which are the main subjects of the 
book. In this chapter are formulated fundamental problems of the BSP, 
given important definitions and described basic mathematical and 
physical models. Moreover, several potential and promising applications 
are reviewed.

Keywords:   Blind Source Separation (BSS), Blind Source Extraction 
(BSE), Independent Component Analysis (ICA), Multichannel Blind 
Deconvolution (MBD), Basic definitions and models, Applications.  


Chapter 2: Solving a System of Linear Algebraic Equations and Related
Problems

In modern signal and image processing fields like biomedical 
engineering, computer tomography (image reconstruction from 
projections), automatic control, robotics, speech and communication, 
linear parametric estimation, models such as auto-regressive 
moving-average (ARMA) and linear prediction (LP) have been extensively 
utilized. In fact, such models can be mathematically described by an 
overdetermined system of linear algebraic equations. Such systems of 
equations are often contaminated by noise or errors, thus the problem of 
finding an optimal and robust with respect noise solution arises if some 
a priori information about the error is available. On the other hand, 
wide classes of extrapolation, reconstruction, estimation, 
approximation, interpolation and inverse problems can be converted to 
minimum norm problems of solving underdetermined systems of linear 
equations. Generally speaking, in signal processing applications, the 
overdetermined system of linear equations describes filtering, 
enhancement, deconvolution and identification problems, while the 
underdetermined case describes inverse and extrapolation problems. This 
chapter provides a tutorial to the problem of solving large 
overdetermined and underdetermined systems of linear equations, 
especially when there is an uncertainty in parameter values and/or the 
systems are contaminated by noise. A special emphasis is placed in 
on-line fast adaptive and iterative algorithms for arbitrary noise 
statistics. This chapter also gives several illustrative examples that 
demonstrate the characteristics of the developed novel algorithms.


Keywords: Least Squares (LS) problem, Extended Total Least Squares 
(TLS), Data Least Squares
(DLS), Least Absolute Deviation (LAD), 1-norm solution, Solving of 
system of linear equations with non-negativity constraints, Non-negative 
Matrix Factorization (NMF), Regularization, Sparse signal 
representation, Sparse solutions, Minimum Fuel Problem (MFP), Focuss 
algorithms, Amari-Hopfield recurrent neural networks for on-line solutions.



Chapter 3: Principal/Minor Component Analysis and Related Problems

Neural networks with unsupervised learning algorithms organize 
themselves in such a way that they can detect or extract useful 
features, regularities, correlations of data or signals or separate or 
decorrelate some signals with little or no prior knowledge of the 
desired results. Normalized (constrained) Hebbian and anti-Hebbian 
learning rules are simple variants of basic unsupervised learning 
algorithms; in particular, learning algorithms for principal component 
analysis (PCA), singular value decomposition (SVD) and minor component 
analysis (MCA) belong to this class of unsupervised rules.  Recently, 
many efficient and powerful adaptive algorithms have been developed for 
PCA, MCA and SVD and their extensions The main objective of this chapter 
is a derivation and overview of the most important adaptive algorithms.  

Keywords: PCA, MCA, SVD, Subspace methods, Automatic dimensionality 
reduction, AIC and MDL criteria, Power method, Robust PCA, Multistage 
PCA for blind source separation.   



Chapter 4: Blind Decorrelation and Second Order Statistics for Robust
Blind Identification

Temporal, spatial and spatio-temporal decorrelations play important 
roles in signal processing. These techniques are based only on 
second-order statistics (SOS). They are the basis for modern subspace 
methods of spectrum analysis and array processing and often used in a 
preprocessing stage in order to improve convergence properties of 
adaptive systems, to eliminate redundancy or to reduce noise. Spatial 
decorrelation or prewhitening is often considered as a necessary (but 
not sufficient) condition for the stronger stochastic independence 
criteria. After prewhitening, the BSS or ICA tasks usually become 
somewhat easier and well-posed (less ill-conditioned), because the 
subsequent separating (unmixing) system is described by an orthogonal 
matrix for real-valued signals and a unitary matrix for complex-valued 
signals and weights. Furthermore, spatio-temporal and time-delayed 
decorrelation can be used to identify the mixing matrix and perform 
blind source separation of colored sources. In this chapter, we discuss 
and analyze a number of efficient and robust adaptive and batch 
algorithms for spatial whitening, orthogonalization, spatio-temporal and 
time-delayed blind decorrelation. Moreover, we discuss several promising 
robust algorithms for blind identification and blind source separation 
of non-stationary and/or colored sources.

Keywords: Robust whitening, Robust orthogonalization, Gram-Schmidt 
orthogonalization,, Second order statistics (SOS) blind identification, 
Multistage EVD/SVD for BSS, Simultaneous diagonalization, Joint 
approximative diagonalization, SOBI and JADE algorithms, Blind source 
separation for non-stationary signals, Natural gradient, Atick-Redlich 
formula, Gradient descent with Frobenius norm constraint.


Chapter 5: Sequential Blind Signal Extraction 

There are three main objectives of this chapter:

(a) To present   simple neural networks (processing units) and propose 
unconstrained extraction and deflation criteria  that do not require 
either  a priori knowledge of source signals or the whitening of mixed 
signals. These criteria lead to simple, efficient, purely local and 
biologically plausible learning rules (e.g., Hebbian/anti-Hebbian type 
learning algorithms).

(b) To prove that the proposed criteria have no spurious equilibriums. 
In other words, the most learning rules discussed in this chapter always 
reach desired solutions, regardless of initial conditions (see 
appendixes for proof).

(c) To demonstrate with computer simulations the validity and high 
performance for practical use of the derived learning algorithms.

In this chapter there are used two different models and approaches. The 
first approach is based on higher order statistics (HOS) which assume 
that sources are mutually statistically independent and they are 
non-Gaussian (expect at most one) and as criteria of independence, we 
will use some measures of non-Gaussianity. The second approach based on 
the second order statistics (SOS) assumes that source signals have some 
temporal structure, i.e., the sources are colored with different 
autocorrelation functions or equivalently different shape spectra. 
Special emphasis will be given to blind source extraction (BSE) in the 
case when sensor signals are corrupted by additive noise using the bank 
of band pass filters.

Keywords: Basic criteria for blind source extraction, Kurtosis, Gray 
function, Cascade neural network, Deflation procedures, KuickNet, 
Fixed-point algorithms, Blind extraction with reference signal, Linear 
predictor and band-pass filters for BSS, Statistical analysis, Log 
likelihood, Extraction of sources from convolutive mixture, Stability, 
Global convergence. 



 Chapter 6: Natural Gradient Approach to Independent Component Analysis

In this chapter, fundamental signal processing and information theoretic 
approaches are presented together with learning algorithms for the 
problem of adaptive blind source separation (BSS) and Independent 
Component Analysis (ICA). We discuss recent developments of adaptive 
learning algorithms based on the natural gradient approach in the 
general linear, orthogonal and Stiefel manifolds. Mutual information, 
Kullback-Leibler divergence, and several promising schemes are discussed 
and reviewed in this chapter, especially for signals with various 
unknown distributions and unknown number of sources. Emphasis is given 
to an information-theoretical and information-geometrical unifying 
approach, adaptive filtering models and associated on-line adaptive 
nonlinear learning algorithms. We discuss the optimal choice of 
nonlinear activation functions for various distributions, e.g., 
Gaussian, Laplacian, impulsive and uniformly-distributed signals based 
on a generalized-Gaussian-distributed model. Furthermore, families of 
efficient and flexible algorithms that exploit non-stationarity of 
signals are also derived.  

Keywords: Kullback-Leibler divergence, Natural gradient concept, 
Derivation and analysis of
natural gradient algorithms, Local stability analysis, Nonholonomic 
constraints, Generalized Gaussian and Cauchy distributions, Pearson 
model. Natural gradient algorithms for non-stationary sources. 
Extraction of arbitrary group of sources, Semi-orthogonality 
constraints, Stiefel manifolds.   



Chapter 7: Locally Adaptive Algorithms for ICA and their Implementations

The main purpose of this chapter is to describe and overview models and 
to present a family of practical and efficient associated adaptive or 
locally adaptive learning algorithms which have special advantages of 
efficiency and/or simplicity and straightforward electronic 
implementations. Some of the described algorithms have special 
advantages in the cases of noisy, badly scaled or ill-conditioned 
signals. The developed algorithms are extended for the case when the 
number of sources and their statistics are unknown.  Finally, problem of 
an optimal choice of nonlinear activation function and general local 
stability conditions are also discussed. In particular, we focus on 
simple locally adaptive Hebbian/anti-Hebbian learning algorithms and 
their implementations using multi-layer neural networks are proposed.

Keywords: Modified Jutten-Herault algorithm, robust local algorithms for 
ICA/BSS, Multi-layer network for ICA, Flexible ICA for unknown number of 
sources, Generalized EASI algorithms, and Generalized stability conditions.



 Chapter 8: Robust Techniques for BSS and ICA with Noisy Data

In this chapter we focus mainly on approaches to blind separation of 
sources when the measured signals are contaminated by large additive 
noise. We extend existing adaptive algorithms with equivariant 
properties in order to considerably reduce the bias caused by 
measurement noise for the estimation of mixing and separating matrices. 
Moreover, we propose dynamical recurrent neural networks for 
simultaneous estimation of the unknown mixing matrix, source signals and 
reduction of noise in the extracted output signals. The optimal choice 
of nonlinear activation functions for various noise distributions 
assuming a generalized-Gaussian-distributed noise model is also 
discussed. Computer simulations of selected techniques are provided that 
confirms their usefulness and good performance.  The main objective of 
this chapter is to present several approaches and derive learning 
algorithms that are more robust with respect to noise than the 
techniques described in the previous chapters or that can reduce the 
noise in the estimated output vector of independent components

Keywords: Bias removal techniques, Wiener filters with references 
convolutive noise, Noise cancellation and reduction, Cumulants based 
cost functions and equivariant algorithms, Blind source separation with 
more sensors than sources, Robust extraction of arbitrary group of 
sources, Recurrent neural network for noisy data, Amari-Hopfield neural 
network.



 Chapter 9: Multichannel Blind Deconvolution:  Natural Gradient Approach

The main objective of this chapter is to review and extend existing 
adaptive natural gradient algorithms for various multichannel blind 
deconvolution models.  Blind separation/deconvolution of source signals 
has been a subject under consideration for more than two decades. There 
are significant potential applications of blind separation/deconvolution 
in various fields, for example, wireless telecommunication systems, 
sonar and radar systems, audio and acoustics, image enhancement and 
biomedical signal processing (EEG/MEG signals). In these applications, 
single or multiple unknown but independent temporal signals propagate 
through a mixing and filtering medium. The blind source 
separation/deconvolution problem is concerned with recovering 
independent sources from sensor outputs without assuming any a priori 
knowledge of the original signals, except certain statistical features.  
In this chapter, we present using various models and assumptions, 
relatively simple and efficient, adaptive and batch algorithms for blind 
deconvolution and equalization for single-input/multiple-output (SIMO) 
and multiple-input/multiple-output (MIMO) dynamical minimum phase and 
non-minimum phase systems. The basic relationships between standard 
ICA/BSS (Independent Component Analysis and Blind Source Separation) and 
multichannel blind deconvolution are discussed in detail.   They enable 
us to extend algorithms derived in the previous chapters. In particular, 
the natural gradient approaches for instantaneous mixture to convolutive 
dynamical models. We also derive a family of equivariant algorithms and 
analyze their stability and convergence properties. Furthermore, a Lie 
group and Riemannian metric are introduced on the manifold of FIR 
filters and using the isometry of the Riemannian metric, the natural 
gradient on the FIR manifold is described. Based on the minimization of 
mutual information, we present then a natural gradient algorithm for the 
causal minimum phase finite impulse response (FIR) multichannel filter. 
Using information back-propagation, we also discuss an efficient 
implementation of the learning algorithm for the non-causal FIR filters. 
Computer simulations are also presented to illustrate the validity and 
good learning performance of the described algorithms.

Keywords: Basic models for blind equalization and multichannel 
deconvolution,  Fractionally Sampled system, SIMO and MIMO models, 
Equalization criteria, Separation-deconvolution criteria, Relationships 
between BSS/ICA and multichannel blind deconvolution (MBD), Natural 
gradient algorithms for MBD, Information Back-propagation.



Chapter 10: Estimating Functions and Superefficiency for ICA and
Deconvolution

Chapter 10 introduces the method of estimating functions to elucidate 
the common structures in most of the ICA/BSS and MBD algorithms. We use 
information geometry for this purpose, and define estimating functions 
in semiparametric statistical models which include unknown functions as 
parameters. Differences in most existing algorithms are only in the 
choices of estimating functions. We then give error analysis and 
stability analysis in terms of estimating functions.  This makes it 
possible to design various adaptive methods for choosing unknown 
parameters included in estimating functions, which control accuracy and 
stability. The Newton method is automatically derived by the 
standardized estimating functions. First the standard BSS/ICA problem is 
formulated in the framework of the semiparametric model and a family of 
estimating functions. Furthermore, the present chapter will discuss and 
extend further convergence and efficiency of the batch estimator  and 
natural gradient learning for blind separation/deconvolution via the   
semiparametric statistical model and estimating functions and 
standardized estimating functions derived by using  efficient score 
functions elucidated  recently by Amari et al. We present the 
geometrical properties of the manifold of the FIR filters based on the 
Lie group structure and formulate the multichannel blind deconvolution 
problem within the framework of the semiparametric model deriving a 
family of estimating functions for blind deconvolution. We then analyze 
the efficiency of the batch estimator based on estimating function - 
obtaining its convergence rate. Finally, we show that both batch 
learning and on-line natural gradient learning are superefficient under 
given nonsingular conditions.

Keywords: Estimating functions, Semiparametric statistical models, 
Superefficiency, Likelihood, Score functions, Batch estimator, 
Information geometry, Stability analysis.



  Chapter 11: Blind Filtering and Separation Using a State-Space Approach

The state-space description of dynamical systems is a powerful and 
flexible generalized model for blind separation and deconvolution or 
more generally for filtering and separation. There are several reasons 
why the state-space models are advantageous for blind separation and 
filtering. Although transfer function models in the Z -domain or the 
frequency domain are equivalent to the state-space models in the time 
domain for any linear, stable time-invariant dynamical system, using 
transfer function directly it is difficult to exploit internal 
representation of real dynamical systems. The main advantage of the 
state-space description is that it not only gives the internal 
description of a system, but there are various equivalent canonical 
types of state-space realizations for a system, such as balanced 
realization and observable canonical forms. In particular, it is 
possible to parameterize some specific classes of models which are of 
interest in applications. In addition, it is relatively easy to tackle 
the stability problem of state-space systems using the Kalman filter. 
Moreover, the state-space model enables a much more general description 
than the standard finite impulse response (FIR) convolutive filtering 
models discussed in the Chapter 9. In fact, all the known filtering 
models, such as the AR, MA, ARMA, ARMAX and Gamma filtering, could also 
be considered as special cases of flexible state-space models. In this 
chapter, we briefly review adaptive learning algorithms based on the 
natural gradient approach and give some perspective and new insight into 
multiple-input multiple-output blind separation and filtering in the 
state-space framework.

Keywords: Linear basic state space model, Natural gradient algorithm for 
state space model, Estimation of output and state space matrices, 
Comparison of various algorithms, Kalman filter, Two stage blind 
separation/filtering approach. 



Chapter 12: Nonlinear State Space Models - Semi-Blind Signal Processing

In this chapter we attempt to extend and generalize the results 
discussed in the previous chapters to nonlinear dynamical models. 
However, the problem is not only very challenging but intractable in the 
general case without    a priori knowledge about the mixing and 
filtering nonlinear process. Therefore, in this chapter we consider very 
briefly only some simplified nonlinear models. In addition, we assume 
that some information about the mixing and separating system and source 
signals is available.  In practice, special nonlinear dynamical models 
are considered in order to simplify the problem and solve it efficiently 
for specific applications. Specific examples include the Wiener model, 
the Hammerstein model and Nonlinear Autoregressive Moving Average models.

Keywords: Semi-blind separation and filtering, Wiener and Hammerstein 
models, Nonlinear Autoregressive Moving Average (NARMA) model, Hyper 
radial basis function (HRBF) neural network.


  Appendix A: Mathematical Preliminaries

In this appendix some mathematical background needed for complete 
understanding of the text are quickly reviewed. Many useful definitions, 
formulas for matrix algebra and matrix differentiation are given

    Keywords:    Matrix inverse update rules, Matrix differentiation,
    Differentiations of scalar cost function with respect to a vector,
    Trace, Matrix differentiation of trace of matrices, Matrix
    expectation, Properties of determinant, Moore-Penrose pseudo
    inverse, Discrimination measures, Distance measures .


 Appendix B:  Glossary of Symbols and Abbreviations
Appendix B contains the list of basic symbols, notation and 
abbreviations used in the book


REFERENCES

The list of references contains more than 1350 publications.



CD-ROM

Accompanying CD-ROM includes electronic, interactive version of the book 
with hyperlinks, full-color figures and text. The black and white 
electronic version with hyperlinks is also provided.
In addition MATLAB user friendly demo package for performing family of 
ICA and BSS/BSE algorithms is included.














More information about the Connectionists mailing list