blind separation of nonlinear mixing signals using … ijvipns-ijens.pdfestimating an orthonormal...

6
International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 27 I J E N S December 2009 IJENS IJENS © - IJVIPNS 7676 - 96310 Blind Separation of Nonlinear Mixing Signals Using Kernel with Slow Feature Analysis Usama S. Mohmmed *1 , Hany Saber *2 1 Department of Electrical Engineering, Assiut University, Assiut 71516, Egypt 2 Department, of Electrical Engineering, South Valley University, Aswan, Egypt 1 Corresponding author [email protected] Abstract -- This paper describes a hybrid blind source separation approach (HBSSA) for nonlinear mixing model (NL- BSS). The proposed hybrid scheme combines simply the kernel- feature spaces separation technique (KTDSEP) and the principle of the slow feature analysis (SFA). The nonlinear mixed data is mapped to high dimensional feature space using kernel-based method. Then, the linear blind source separation (BSS) based on the slow feature analysis (SFA) is used to extract the most slowness vectors among the independent data vectors. The proposed scheme is based on the following four key features: 1) estimating an orthonormal bases, 2) mapping the data into the subspace using this orthonormal bases, 3) applying linear BSS on the mapping data to make the data vectors in the feature spaces are independent, 4) Applying the principle of slow feature analysis on the mapping data to select the desired signals. The SFA provides the dimension reduction according to the most independent and slowing variable signals. Moreover, the orthonormal bases estimation in the wavelet domain is introduced in this work to reduce the complexity of the KTDSEP algorithm. The motivation of using the wavelet transform, in estimating the orthonormal bases, is based on the fact that the low frequency band in the wavelet domain contains the significant power of the signal. The advantages of the proposed method are the fast estimation of the orthonormal bases and the dimension reduction of the estimating data vectors. Performed computer simulations have shown the effectiveness of the idea, even in presence of strong nonlinearities and synthetic mixture of real world data. Our extensive experiments have confirmed that the proposed procedure provides promising results . Index Term-- Nonlinear blind source separation, slow feature analysis, independent slow feature analysis, slow feature analysis, kernel base algorithm, and kernel trick I. INTRODUCTION The problem of blind source separation (BSS) consists on the recovery of independent sources from their mixture. This is important in several applications like speech enhancement, telecommunication, biomedical signal processing, etc. Most of the work on BSS mainly addresses the cases of instantaneous linear mixture [1]-[5]. Let A is a real or complex rectangular (n×m; n≥m) matrix, the data model for linear mixture can be expressed as follows: AS(t) X(t) (1) Where S(t) represents the statistically independent sources array while X(t) is the array containing the observed signals. For real world situation, however, the basic linear mixing model in equation (1) is too simple for describing the observed data. In many applications such as the nonlinear characteristic introduced by preamplifiers of receiving sensors, we can consider a nonlinear mixing. Therefore, a nonlinear mixing is more realistic and accurate than linear model. For instantaneous mixtures, a general nonlinear data model can have the form: (S(t)) X(t) f (2) Where f is an unknown vector of real functions. The linear instantaneous mixing (BSS) can be solved by using independent component analysis (ICA) [4]. The goal of the ICA is to separate the signals by finding independent component from the data signal. It is important to note that if x and y are two independent random variables, any of their functions f (x) and f (y) are also independent. An even more serious problem is that in the nonlinear case, x and y can be mixed and still be statistically independent. Several authors [6-10] have addressed the important issues on the existence and uniqueness of solutions for the nonlinear ICA and BSS problems. In general, the ICA is not a strong enough constraint for ensuring separation in the nonlinear mixing case. There are several known methods that try to solve this nonlinear BSS problem. They can roughly be divided into algorithms with a parametric approach and algorithms with nonlinear expansion approach. With the parametric model the nonlinearity of the mixture is estimated by parameterized nonlinearities. In [8] and [11], neural network is used to solve this problem. In the nonlinear expansion approach the observed mixture is mapped into a high dimensional feature space and afterwards a linear method is applied to the expanded data. A common technique to turn a nonlinear problem into a linear one is introduced in [12]. I.1 Nonlinear mixture model A generic nonlinear mixture model for blind source separation can be described as follows: (S(t)) f A X(t) (3) Where S(t) represents the statistically independent sources array while X(t) is the array containing the observed signals and f is unknown multiple-input and multiple-output (MIMO) mapping which called the nonlinear mixing transform (NMT). In order for the mapping to be invertible, we assume that the nonlinear mapping is monotone. We make the assumption here, for simplicity and convenience, that the dimensions of X and S is equal. An important special case of the nonlinear mixture is the so- called post-nonlinear (PNL) mixture (AS(t)) X(t) f (4) Where f is an invertible nonlinear function that operates componentwise and A is a linear mixing matrix, more detailed n 1,....., i , (t) S a f (t) X m 1 j j ij i i (5) Where a ij are the elements of the mixing matrix A. The PNL model was introduced by Taleb and Jutten[13]. It represents an important subclass of the general nonlinear model and has therefore attracted the interest of several researchers [14]-[18].

Upload: hoangnhi

Post on 24-May-2018

218 views

Category:

Documents


2 download

TRANSCRIPT

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 27

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

Blind Separation of Nonlinear Mixing Signals Using

Kernel with Slow Feature Analysis Usama S. Mohmmed

*1, Hany Saber

*2

1Department of Electrical Engineering, Assiut University, Assiut 71516, Egypt

2 Department, of Electrical Engineering, South Valley University, Aswan, Egypt

1 Corresponding author [email protected]

Abstract-- This paper describes a hybrid blind source

separation approach (HBSSA) for nonlinear mixing model (NL-

BSS). The proposed hybrid scheme combines simply the kernel-

feature spaces separation technique (KTDSEP) and the principle

of the slow feature analysis (SFA). The nonlinear mixed data is

mapped to high dimensional feature space using kernel -based method. Then, the linear blind source separation (BSS) based on

the slow feature analysis (SFA) is used to extract the most

slowness vectors among the independent data vectors. The

proposed scheme is based on the following four key features: 1)

estimating an orthonormal bases, 2) mapping the data into the subspace using this orthonormal bases, 3) applying linear BSS on

the mapping data to make the data vectors in the feature spaces

are independent, 4) Applying the principle of slow feature

analysis on the mapping data to select the desired signals. The

SFA provides the dimension reduction according to the most independent and slowing variable signals. Moreover, the

orthonormal bases estimation in the wavelet domain is

introduced in this work to reduce the complexity of the KTDSEP

algorithm. The motivation of using the wavelet transform, in

estimating the orthonormal bases, is based on the fact that the low frequency band in the wavelet domain contains the

significant power of the signal. The advantages of the proposed

method are the fast estimation of the orthonormal bases and the

dimension reduction of the estimating data vectors. Performed

computer simulations have shown the effectiveness of the idea, even in presence of strong nonlinearities and synthetic mixture of

real world data. Our extensive experiments have confirmed that

the proposed procedure provides promising results.

Index Term-- Nonlinear blind source separation, slow feature

analysis, independent slow feature analysis, slow feature analysis,

kernel base algorithm, and kernel trick

I. INTRODUCTION

The problem of blind source separation (BSS) consists on the

recovery of independent sources from their mixture. This is

important in several applications like speech enhancement,

telecommunication, biomedical signal processing, etc. Most of

the work on BSS mainly addresses the cases of instantaneous

linear mixture [1]-[5]. Let A is a real or complex rectangular

(n×m; n≥m) matrix, the data model for linear mixture can be

expressed as follows:

AS(t)X(t) (1)

Where S(t) represents the statis tically independent sources

array while X(t) is the array containing the observed signals. For real world situation, however, the basic linear mixing

model in equation (1) is too simple for describ ing the

observed data. In many applications such as the nonlinear

characteristic introduced by preamplifiers of receiving

sensors, we can consider a nonlinear mixing. Therefore, a

nonlinear mixing is more realistic and accurate than linear

model. For instantaneous mixtures, a general nonlinear data

model can have the form:

(S(t))X(t) f (2)

Where f is an unknown vector of real functions. The linear

instantaneous mixing (BSS) can be solved by using

independent component analysis (ICA) [4]. The goal of the

ICA is to separate the signals by finding independent

component from the data signal. It is important to note that if

x and y are two independent random variables, any of their

functions f(x) and f(y) are also independent. An even more

serious problem is that in the nonlinear case, x and y can be

mixed and still be statistically independent. Several authors

[6-10] have addressed the important issues on the existence

and uniqueness of solutions for the nonlinear ICA and BSS

problems. In general, the ICA is not a strong enough

constraint for ensuring separation in the nonlinear mixing

case. There are several known methods that try to solve this

nonlinear BSS problem. They can roughly be divided into

algorithms with a parametric approach and algorithms with

nonlinear expansion approach. With the parametric model the

nonlinearity of the mixture is estimated by parameterized

nonlinearities. In [8] and [11], neural network is used to solve

this problem. In the nonlinear expansion approach the

observed mixture is mapped into a high dimensional feature

space and afterwards a linear method is applied to the

expanded data. A common technique to turn a nonlinear

problem into a linear one is introduced in [12].

I.1 Nonlinear mixture model

A generic nonlinear mixture model for blind source separation

can be described as follows:

(S(t)) f A X(t)

(3)

Where S(t) represents the statistically independent sources

array while X(t) is the array containing the observed signals

and f is unknown multiple-input and multiple-output (MIMO)

mapping which called the nonlinear mixing transform (NMT).

In order for the mapping to be invertible, we assume that the nonlinear mapping is monotone. We make the assumption

here, for simplicity and convenience, that the dimensions of X

and S is equal.

An important special case of the nonlinear mixture is the so-

called post-nonlinear (PNL) mixture

(AS(t)) X(t) f (4)

Where f is an invertible nonlinear function that operates

componentwise and A is a linear mixing matrix, more detailed

n1,.....,i,(t)S af(t)Xm

1j

jijii

(5)

Where aij are the elements of the mixing matrix A. The PNL

model was introduced by Taleb and Jutten[13]. It represents

an important subclass of the general nonlinear model and has

therefore attracted the interest of several researchers [14]-[18].

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 28

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

Applications are found, for example, in the fields of

telecommunications, where power efficient wireless

communication devices with nonlinear class C amplifiers are

used [19] or in the field of biomedical data recording, where

sensors can have nonlinear characteristics [20].

In this work, a new method to solve the PNL-BSS problem or

the general nonlinear BSS is proposed. In the first step, the

nonlinear mixed data is mapped to higher dimensional feature

space fk and linear blind source separation algorithm is applied

on the mapped data. In the second step, the principle of slow

feature analysis is used on the linearly separated data to find

the most slow and independent vectors in this feature space. In

the following Sections, the kernel-feature Spaces separation

technique (KTDSEP) and the principle of slow feature

analysis (SFA) will be discussed.

II. KERNEL FEATURE SPACES AND NONLINEAR BLIND SOURCE

SEPARATION

In kernel-based learning, the data is mapped to a kernel

feature space of a dimension that corresponds to the number

of training data points. Suppose that the data is Xi

(i=1,………..,N), so the idea is to mapping the data Xi into

some kernel feature space ƒk by some mapping Φ : Rn→ ƒk.

Performing a simple linear algorithm in ƒk, corresponds to a

nonlinear algorithm in the input space will solve the

separation problem. Essential ingredients to kernel based

learning are: (1) support vector machine (SVM) [21] [22]

theory that can provide a relation between the complexity of

the function class in use and the generalization error, and (2)

the famous kernel trick

Φ(y)Φ(x)y)K(x,

(6)

Which efficiently computes the scalar product. This trick is essential if ƒk is an infinite dimensional space. Even though ƒk

might be infinite dimensional the subspace, where the data lies

is maximally N-dimensional. However, the data typically forms an even smaller subspace in ƒk [23]. To map the

nonlinear mixing data X1… XN nR into the feature space

ƒk the kernel trick in Equation (6) must be used, but the high

domination mapping data will be. In [21] the idea of estimating the orthonormal bases for subspace ƒk is proposed

and these bases is used to map the nonlinear mixed data to the

feature space ƒk

For some further points v1,……..,vd nR from the same

space, that will later generate a bases in ƒk. The mapping

method will be as follows:

Consider the mapping points as )(....)( 1 Nx xx

,

and )(....)( 1 dv vv

.

Assume that the columns of Φv constitute a base of the column

space of Φx, so

)( and )()( drankspanspan v

T

vxv

(7)

Moreover, Φv being bases implies that the matrix ΦvT Φv has

full rank and its inverse exists. Now an orthonormal bases can

be defined as follows:

vvv2

1

)(

(8)

The column space of which is identical to the column space of

Φv. Consequently, the bases enable us to parameterize all

vectors that lie in the column space of Φx by some vectors

indR . The space that is spanned by is called parameter

space. The orthonormal bases Equation (7) enables the

working in dR the span of , which is extremely valuable

since d depends solely on the kernel function and the

dimensionality of the input space. Therefore, d is independent

of N.

To map the mixed data from input space to feature space fk

the kernel trick is used to perform the following:

dji

vvkvv jijiijvv

........1, with

),()()()(

Njdi

xvkxv jijiijxv

........1,........1, with

),()()()(

(9)

Finally the mapping matrix will be

xvvvxx

2

1

)(

(10)

Which is also a real valued d x N matrix and the matrix

2

1

)(

vv is symmetric. Two main problems facing the

kernel based algorithm (KTDSEP):

Selecting the points v1,…..,vd to constructed the

orthonormal bases

The number of output signals from KTDSEP algorithm will

be d signals and we need only n signals where n is smaller

than d.

In the traditional KTDSEP algorithms, the random sampling

method and the k-mean clustering method are used to find the

orthonormal bases. These two techniques increase the

complexity of the algorithm. Moreover, the KTDSEP method

cannot automatically find n signals out of d signals. The

proposed solution by the KTDSEP method [24] is based on

the correlation among the original signals and the output

signals. In the blind source separation field the original signals

is not available, so this technique is not practical. Another

proposed solution [21] is done by applying the KTDSEP on

the signals two times. This solution will increase the

computational complexity of the algorithm.

In this work, after mapping the data, the linear blind source

separation algorithm is used. The temporal decorrelation BSS

(TDSEP) [24] is used as linear blind source separation

algorithm. In the next Sections, the details of our proposed

solution will be discussed.

III. SLOW FEATURE ANALYSIS AND NONLINEAR BLIND

SOURCE SEPARATION

Slow Feature Analysis (SFA) is a method that extracts slowly

varying signals from a given observed signals [25], [26]. Consider an input signal x(t)=[x1(t), . . . ,xN(t)]

T, the objective

of SFA is to find a nonlinear input-output function g(x) =

[g1(x), . . . , gL(x)]T such that the components of u(t)= g(x(t))

are varying as slowly as possible. The variance of the first

derivative is used as a measurement of the slowness. The

optimization problem will be as follows:

Minimize the objective function

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 29

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

)())((2

'

tUtU ii

(11)

Successively for each Ui(t) under the following

constraints:

0)( tui

(a)

1)(2tui

(b)

0)()( ijtutu ji

(c)

Where <•>denotes averaging over time. Constraints (a), (b)

and (c) ensure that the solution will not be the trivial solution

(Ui(t)= const). Constraint (c) provides uncorrelated output

signal components and thus guarantees that different

components carry different information.

Note that slowly varying signal components are easier to

predict and should therefore have strong correlations in time.

In the proposed work, the principle of slow feature analysis is

used to overcome the main drawback of the kernel base

algorithm (KTDSEP) to pickup n signals out of d signals. The

slow feature analysis algorithm is motivated by the fact that

the linear signals are the slowest varying signals among all the

signals [12]. Therefore, the slow feature analysis can be used

to pickup n slow varying signals (linear) from d signals.

The slow feature analysis algorithm can be summarized in the

following steps:

i. Nonlinear expanding the input data.

ii. The nonlinear expanding signal is whitened using

eigenvalue decomposition of the zero time-log correlation

matrixes.

iii. Derivative the whitened signal is calculated according to

)()1()(y '

tytyt

iv. The rotating matrix Q is calculated using the joint

approximate diagonalization (JAD).

v. The eigenvalue and the eigenvectors of Q are evaluated.

vi. The dimensional reduction of the signals by is achieved

using the normalized eigenvectors that corresponding to the

smallest eigenvalue of the rotation matrix Q.

IV. THE PROPOSED HYBRID BLIND SOURCE SEPARATION

APPROACH (HBSSA)

One of the most popular methods to solve the problem of the

nonlinear mixing model (NL-BSS) is the mapping of the

nonlinear mixing data into high dimensional feature space fk.

Then the problem can be handled as a linear problem under

some constraints [12], [21]. As mentioned in Section II, the

kernel base algorithm (KTDSEP) is depended on mapping the

nonlinear mixed data into high dimensional space.

The proposed separation method combines simply the kernel-

feature Spaces separation technique (KTDSEP) and the

principle of the slow feature analysis (SFA). The estimation of

the orthonormal bases is done in the wavelet domain to reduce

the complexity of the KTDSEP algorithm. The motivation of

using the wavelet transform, in estimating the orthonormal

bases, is based on the fact that the low frequency band in the

wavelet domain contains the significant power of the signal.

Moreover, the SFA is used to select the desired signals from

the output signals of the KTDSEP algorithm. The SFA

algorithm is motivated by the fact that the linear signals are

the slowest varying signals among all the signals. The

orthonormal bases is started by selecting the data vector

v1,………….., vd and testing these vector using Equation (6).

This vector is selected using the k-mean clustering in KTDSEP. The wavelet transform is applied on the nonlinear

mixed data n-times to select the v1,………….., vd from lowest

frequency band. Then, equation (7) is used to find the

orthonormal bases . The proposed method can be

summarized as the follows:

1. The wavelet transformed (using wavelet packet) is

applied on the nonlinear mixed data n times.

2. The points v1,….., vd are selected from the lowest

frequency band in the wavelet domain.

3. Equation (7) is used to construct the orthonormal

bases .

4. The kernel t rick is used to map the mixed data into the nonlinear feature space fk and find d x N matrix

5. The output data ψx are pre-whitened

6. Evaluate n time shifted covariance matrices among the

vectors of pre-whitened ψx

7. The joint approximate diagonilaizat ion criteria is used

with the n time shifted covariance matrices to estimate the

d x d separating matrix B.

8. The Slow Feature Analysis is applied on the d x d

separation matrix B to get n x d rotation matrix (QASF) as

following:

Derivative the whitened signals ψx step (5)

The rotating matrix Q is calculated using the joint

approximate diagonalization (JAD).

The eigenvalue and the eigenvectors of Q are

evaluated

Applying the dimensional reduction to find n out

of d signals by using the n normalized eigenvectors

that corresponding to the smallest n eigenvalue of the

rotation matrix Q and find QASF= Q(mxd) x B(dxd)

9. Multiply the QASF by the output d x N, from step (5), to

get n x d output signals.

Steps 5, 6 and 7 are refer to the TDSEP algorithm and step 8 is

the principle of SFA algorithm

V. ASSUMPTIONS AND DEFINITIONS

In the proposed algorithm, there are some assumptions are

considered as follows: (1) the source signals must be

independent, (2) the number of mixed signals are greater than

or equal to the number of the source signals, (3) there are at

most one source signal have Gaussian distribution, (4) the

nonlinear function f is invertible function, and (5) the mixed

matrix must be of full rank.

VI. SIMULATION RESULTS AND IMPLEMENTATION

Several types of signals are used to evaluate the performance

of the proposed technique. To put the results in a comparison

form, the same examples in [21] and we introduced a new

example.

Experiment 1: In the first experiment, two sinusoidal signals

with different frequencies are used: T

21 (T)]S (t)[SS(t)

where )100sin()(1 ttS

and )42sin()(2 ttS

,

)( 2

1

xvvvxx

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 30

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

with2000,.......,1t

. These source signals are nonlinearly

mixed according to the following equation:

)()(

2

)()(

1

21

21

)(

)(

tsts

tsts

eetX

eetX

(12)

Here, a polynomial kernel of degree 5 is used,

5)1(),( babak T

(13)

In this experiment, 21 points (v1,….,v21) are selected. So we

have ψx with size of 21 x 2000. When TDSEP is applied on ψx, 21 different data vectors are generated and the slow

feature analysis generates 2 x 2000 data vectors. Fig. 1 shows

the original, the mixing and the estimating sources. Moreover,

the scattering plot of the three signals is shown in Fig 2.

Experiment 2: In this experiment, two speech signals, T

21 (T)]S (t)[SS(t) , of 2000 samples are mixed. The

nonlinear mixing is done using Equation (14). Fig. 3 shows

the original, the mixing and the estimating sources. The

scattering plot of the original signals, the mixed signals and

the estimating signals are shown in Fig 4.

))(sin()1)((5.1)(

))(cos()1)(()(

122

121

tStStX

tStStX

(14)

Gaussian radial base function (RBF) Equation (15) is

employed with σ = 0.5.

e σ

ba

k(a,b) 2

2

(15)

In this experiment, 17 points (v1,….,v17) is selected so the ψx

has a size of 17 x 2000 is estimated. When TDSEP is applied

on ψx we will have 17 different data vectors. The slow feature

analysis is used to get 2 x 2000 data vectors from 17 x 2000

different data vectors.

Fig. 1. the original data are the first two signals on the top, the mixing data are

in the middle and the recovering data are in the last two rows

(a) (b) ( c)

Fig. 2. Scatter plot of: (a) the original data (b) the mixed data. (c) the estimated

data

Experiment 3: In this experiment, the same speech signals

in experiment 2 are used. The nonlinear twisted function [12],

Equation (16), is used. Fig. 5 shows the original, the mixing

and the estimating sources. The scattering plot of the original

signals, the mixed signals and the estimating signals are

shown in Fig 6.

))(5.1sin()6)(3)(()(

))(5.1cos()6)(3)(()(

1122

1121

tStStStX

tStStStX

(16)

The same RBF in Equation (15) with σ = 0.5 is employed in this experiment. 20 points (v1,….,v20) is selected so we will get

ψx of size 20 x 2000. When TDSEP is applied on ψx we will

have 20 different data vectors. The slow feature analysis is

used to get 2 x 2000 data vectors from 20x2000 different data

vectors.

Fig. 3. the original data are the first two signals on the top, the mixing

data are in the middle and the recovering data are in the last two rows

(a) (b) ( c)

Fig. 4. (a) scatter plot of the original data. (b) Scatter plot of the mixed data.(c) the estimated data

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 31

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

Fig. 5. the original data are the first two signals on the top, the mixing data

are in the middle and the recovering data are in the last two rows

(a) (b) ( c)

Fig. 6. (a) scatter plot of the original data. (b) Scatter plot of the mixed data. (c) scatter plot of the estimated data

Experiment 4: In this experiment two speech signals with

length 30000 samples T

21 (T)]S (t)[SS(t) are used, the

nonlinear mixed is done using Equation (17). Fig. (7) shows

the original, the mixing and the estimating sources. Moreover,

the scattering plot of the original signals, the mixed signals

and the estimating signals are shown in Fig. (8).

))]()((sin[)(

)]()(tanh[)(

122

121

tStStX

tStStX

(17)

In this experiment a polynomial kernel equation (23) is used

as a kernel function but with degree 7, and 19 point

(v1,….,v19) is selected, so the dimension of ψx will be 19 ×

30000. When linear BSS is applied on ψx 19 different vectors

will be generated. The slow feature analysis is used to get 2 ×

30000 vectors from 19 × 20000 different vectors.

Fig. 7. the original data are the first two signals on the top, the mixing data

are in the middle and the recovering data are in the last two rows

(a) (b) ( c)

Fig. 8. (a) scatter plot of the original data. (b) Scatter plot of the mixed data. (c) scatter plot of the estimated data

VII. CONCLUDING REMARKS

In this paper, a novel hybrid scheme to solve the problem of

blind source separation in nonlinear mixing model (NL-BSS)

is proposed. The proposed hybrid scheme combines simply

the kernel-feature Spaces separation technique (KTDSEP) and

the principle of the slow feature analysis (SFA). The proposed

algorithm overcomes the drawback of the kernel base

nonlinear blind source separation algorithm (KTDSEP). The

slow feature analysis principle is used in selecting the n

separated signals from d output signals. Moreover, new

approach to construct the orthonormal bases by selecting the

bases points from the lowest frequency band in the wavelet

transformed is introduced. The advantage in performance of

HBSSA lays in the complexity reduction in the orthonormal

bases estimation process and the fast selection of the desired

signals from the output signals of the KTDSEP algorithm.

REFERENCES

[1] Amari, A. Cichocki, and H. H. Yang. "A new learning algorithm for blind signal separation." In NIPS 95, pp. 882–893. MIT Press, 1996.

[2] Bell and T . Sejnowski. ―Information-maximization approach to blind separation and blind deconvolution" Neural Computation, vol. 7, pp.1129–1159, 1995.

[3] Cardoso. "infomax and maximum likelihood for source separation" IEEE Letters on Signal Processing, vol. 4, pp.112–114, 1997.

[4] W. Lee, M. Girolami, and T . Sejnowski. "Independent component analysis using an extended infomax algorithm for mixed sub-

Gaussian and super-Gaussian sources." Neural Computation, vol. 11, pp.417–441, 1999.

[5] D.-T . Pham, and J.-F. Cardoso, ―Blind separation of instantaneous mixtures of non-stationary sources‖, IEEE Trans. on Signal

Processing, vol. 49, no.9, pp. 1837-1848, Sep. 2001. [6] C. Jutten and A. Taleb, ―Source separation: from dusk till dawn,‖ in

Proc. 2nd Int. Workshop on Independent Component Analysis and Blind Source Separation (ICA2000), pp. 15–26, Helsinki, Finland,

2000. [7] A. Hyv¨arinen and P. Pajunen, ―Nonlinear independent component

analysis: Existence and uniqueness results,‖ Neural Networks, vol.

12, no. 3, pp. 429–439, 1999. [8] A. Taleb and C. Jutten, ―Source separation in post -nonlinear

mixtures,‖ IEEE Trans. on Signal Processing, vol. 47, no. 10, pp. 2807–2820, 1999.

[9] A. Taleb, ―A generic framework for blind source separation in structured nonlinear models,‖ IEEE Trans. on Signal Processing, vol. 50, no. 8, pp. 1819–1830, 2002.

[10] J. Eriksson and V. Koivunen, ―Blind identifiability of class of

nonlinear instantaneous ICA models,‖ in Proc. of the XI European Signal Proc. Conf. (EUSIPCO 2002), vol. 2, Toulouse, France, , pp. 7–10, September 2002.

[11] Luis B. Almeida, "Linear and nonlinear ICA based on mutual

information‖ the MISEP method, Signal Processing, Special Issue on Independent Component Analysis and Beyond 2004, vol. 84, no.2, pp.231–245.

[12] Tobias Blaschke and Laurenz Wiskott . "Independent Slow Feature

Analysis and Nonlinear Blind Source Separation "Proc. of the 5th Int. Conf. on Independent Component Analysis and Blind Signal Separation Granada, Spain, pp. 22-24 September 2004

International Journal of Video& Image Processing and Network Security IJVIPNS-IJENS Vol:09 No:10 32

I J E N S December 2009 IJENS IJENS © -IJVIPNS7676 -96310

[13] Taleb and C. Jutten. Nonlinear source separation: The post -nonlinear mixtures. In Proc. European Symposium on Artificial

Neural Networks, pp. 279–284, Belgium, 1997. [14] T .-W. Lee, B.U. Koehler, and R. Orglmeister. Blind source

separation of nonlinear mixing models. IEEE International Workshop on Neural Networks for Signal Processing, pp. 406–415,

1997. [15] H.-H. Yang, S. Amari, and A. Cichocki. Information-theoretic

approach to blind separation of sources in non-linear mixture. Signal Processing, vol. 64, no.3, pp.291–300, 1998.

[16] A. Taleb and C. Jutten. Source separation in post -nonlinear mixtures. IEEE Trans. on Signal Processing, vol. 47, no.10, pp.2807–2820, 1999.

[17] A. Ziehe, M. Kawanabe, S. Harmeling, and K.-R. M¨uller.

Separation of post -nonlinear mixtures using ACE and temporal decorrelation. In T .-W. Lee, editor, Proc. Int . Workshop on Independent Component Analysis and Blind Signal Separation

(ICA2001), pp. 433–438, San Diego, California, 2001. [18] C. Jutten and J. Karhunen. Advances in nonlinear blind source

separation. In Proc. of the 4th Int. Symp. on Independent Component Analysis and Blind Signal Separation (ICA2003), pp.

245–256, Nara, Japan, Invited paper in the special session on nonlinear ICA and BSS, April 2003.

[19] L. E. Larson. "Radio frequency integrated circuit technology low-power wireless communications". IEEE Personal Communications,

vol. 5, no. 3, pp.11–19, 1998. [20] A. Ziehe, K.-R. M¨uller, G. Nolte, B.-M. Mackert, and G. Curio.

Artifact reduction in magnetoneurography based on time-delayed second-order correlations. IEEE Trans. Biomed. Eng., vol. 47, no.

1, pp. 75–87, 2000. [21] Harmeling, S., Ziehe, A., Kawanabe, M., and Müller, K.-R.

"Kernel-based nonlinear blind source separation." Neural

Computation, vol. 15, pp: 1089–1124, 2003. [22] N. Cristianini and J. Shawe-Taylor.(200) " An Introduction t o

Support Vector Machines." Cambridge University Press, Cambridge, UK, 2000

[23] B. Schölkopf, S. Mika, C.J.C. Burges, P. Knirsch, K.-R. Müller, G. Rätsch, and A.J. Smola." Input space vs. feature space in kernel-based methods." IEEE Transactions on Neural Networks, vol. 10, no. 5 pp.1000–1017, 1999

[24] A. Ziehe and K.-R. Müller. "TDSEP—an efficient algorithm for blind separation using time Structure,‖ In Proc. Int. Conf. on Artificial Neural Networks (ICANN’98), pp. 675–680, 1998

[25] Laurenz Wiskott "Slow feature analysis: a theoretical analysis of

optimal free responses" Neural Computation, vol. 15, pp.2147 - 2177, 2003

[26] Blaschke, T ., Berkes, P., and Wiskott, L. "What is the relation between independent component analysis and slow feature

analysis?" Neural Computation, vol. 18, no.10, pp.2495–2508. 2006