adaptive blind source separation based on intensity vector...

181
Adaptive Blind Source Separation Based on Intensity Vector Statistics A. RIAZ Submitted for the Degree of Doctor of Philosophy from the University of Surrey Centre for Vision, Speech and Signal Processing Faculty of Engineering and Physical Sciences University of Surrey Guildford, Surrey GU2 7XH, U.K. December 2015 © A. RIAZ 2015

Upload: others

Post on 03-Feb-2020

11 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Adaptive Blind Source Separation Based onIntensity Vector Statistics

A. RIAZ

Submitted for the Degree ofDoctor of Philosophy

from theUniversity of Surrey

Centre for Vision, Speech and Signal ProcessingFaculty of Engineering and Physical Sciences

University of SurreyGuildford, Surrey GU2 7XH, U.K.

December 2015

© A. RIAZ 2015

Page 2: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates
Page 3: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Summary

Human brain has the ability to focus on desired acoustic source when several sourcesare active. In the domain of digital electronics this problem is termed as the cock-tail party problem. Over the past few decades many algorithms have been proposedwhich attempt to solve this problem; they are generally termed as acoustic source sep-aration algorithms. The proposed algorithms achieve separation of individual sourcecomponents from observed acoustic mixtures. The source separation system may becapable of estimating the number of sources, their physical locations, the room impulseresponse and/or any target source signal information. A system that approximatesthis information is termed as blind. Source separation systems which require any suchinformation beforehand are termed as semi-blind.

Most of the proposed source separation algorithms deal with acoustic sources that arestationary in space. A more challenging task is to approximate unmixing filters whilethe sources are constantly moving. To maintain output performance in such a scenario,the source separation system has to swiftly and accurately detect the time variant mix-ing parameters, and update unmixing filters accordingly. The area of moving sourceshas still not been heavily investigated by researchers.

The aim of this thesis is to further the field of acoustic source separation. Investigationof intensity vector direction (IVD) based source separation algorithm was carried outto analyse and improve the system, both in terms of applicability and output soundquality. The algorithm under investigation provides a robust and nearly closed-formsolution to the source separation problem with a low processing time. However, the al-gorithm initially required unmixing filter coefficients as input for dealing with practicalacoustic scenarios. Analysis performed with microphone array response, microphonearray geometry and the room response yielded three different modifications to thebaseline system, improving system applicability and output sound quality.

The IVD based system was investigated to deal with more challenging acoustic scenar-ios, such as time variant number of sources. Likewise, the IVD statistics were analysedto propose solutions for moving sources scenario. The system exhibited potential toswiftly, accurately and reliably detect changes in the time varying mixing parameters.As a result of these investigations, a novel system pipeline is proposed, capable ofdetecting, tracking and separating moving sources in a blind manner.

The proposed algorithms were evaluated for processing time and separation perfor-mance. Optimisation of output sound quality was carried out through objective perfor-mance measures, while speaker tracking was evaluated subjectively. Finally, a demon-stration was developed in Matlab based on the proposed algorithms to facilitate userinteraction with the surrounding acoustic environment.

Key words: convolutive mixtures, moving blind source separation, Intensity VectorDirection statistics, low processing time, coincident array processing.

Page 4: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Email: [email protected]

WWW: http://www.eps.surrey.ac.uk/

Page 5: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Acknowledgements

It gives me great pleasure to acknowledge Prof. Ahmet Kondoz, who gave me theopportunity to pursue a doctorate degree. His enthusiasm and knowledge had a lastingeffect and without his guidance, patience and persistent help this thesis would not havebeen possible. I would like to specially thank Prof. Mark Plumbley for his guidancein improving the presentation of my work. I would like to thank my co-supervisorDr. Xiyu Shi for always encouraging me throughout these years, and giving me helpfulreviews. Last but not the least I would like to thank my parents, my sister and mybeloved wife for the never ending love and support.

Page 6: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Contents

Nomenclature xvi

1 Introduction 1

1.1 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2 Scope and Objective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

1.3 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.4 Original Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

1.5 Publications by the Author . . . . . . . . . . . . . . . . . . . . . . . . . 7

2 Related Work 8

2.1 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

2.1.1 Source Mixture Models . . . . . . . . . . . . . . . . . . . . . . . 9

2.1.2 Statistical Methods . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.1.3 Beam-forming Methods . . . . . . . . . . . . . . . . . . . . . . . 17

2.1.4 Sparsity-based Methods . . . . . . . . . . . . . . . . . . . . . . . 19

2.1.5 Moving Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

2.2 Intensity Vector Direction based Source Separation . . . . . . . . . . . . 23

2.2.1 Spatial Audio and B-format . . . . . . . . . . . . . . . . . . . . . 24

2.2.2 B-Format Audio for IVD Measurement . . . . . . . . . . . . . . 25

2.2.3 Microphone System for IVD Measurement . . . . . . . . . . . . . 28

2.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

i

Page 7: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Contents ii

3 Improvement in IVD-based Source Separation for Stationary Sources 30

3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

3.2 Baseline IVD-Based Source Separation System . . . . . . . . . . . . . . 33

3.2.1 Pressure and Pressure Gradient Measurements . . . . . . . . . . 36

3.2.2 Frequency Domain Representation . . . . . . . . . . . . . . . . . 37

3.2.3 A-format to B-format Conversion . . . . . . . . . . . . . . . . . . 37

3.2.4 IVD measurement . . . . . . . . . . . . . . . . . . . . . . . . . . 38

3.2.5 Spatial Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

3.2.6 Time Domain Representation . . . . . . . . . . . . . . . . . . . . 39

3.3 Analysis of the IVD-based System . . . . . . . . . . . . . . . . . . . . . 39

3.3.1 Objective Performance Evaluation . . . . . . . . . . . . . . . . . 40

3.3.2 A-format Array Characteristics . . . . . . . . . . . . . . . . . . . 44

3.3.3 Effect of the Room Environment . . . . . . . . . . . . . . . . . . 53

3.3.4 Interaction Amongst Sources . . . . . . . . . . . . . . . . . . . . 58

3.4 Microphone Array Correction . . . . . . . . . . . . . . . . . . . . . . . . 59

3.4.1 Evaluation of Microphone Array Correction . . . . . . . . . . . . 61

3.5 Cardioid Weight Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

3.6 Adaptive Spatial Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . 67

3.6.1 Rate of Adaptation of Spatial Filter Parameters . . . . . . . . . 72

3.7 Combining All Modifications . . . . . . . . . . . . . . . . . . . . . . . . 74

3.8 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76

3.8.1 Empirical Computational Time . . . . . . . . . . . . . . . . . . . 76

3.8.2 Two Source Mixtures . . . . . . . . . . . . . . . . . . . . . . . . 77

3.8.3 Four Source Mixtures . . . . . . . . . . . . . . . . . . . . . . . . 81

3.9 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85

4 Blind Source Separation of Moving Speakers Based on Intensity Vec-tor Statistics 87

4.1 Preview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

4.2 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88

4.3 Analysis of the IVD Distribution in Space . . . . . . . . . . . . . . . . . 89

4.3.1 Acoustic Source Activity . . . . . . . . . . . . . . . . . . . . . . . 91

Page 8: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Contents iii

4.3.2 Effect of Distance to the Microphone Array . . . . . . . . . . . . 94

4.3.3 Moving Acoustic Sources . . . . . . . . . . . . . . . . . . . . . . 95

4.4 Proposed Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96

4.4.1 IVD Distribution in Space . . . . . . . . . . . . . . . . . . . . . . 97

4.4.2 Peak Detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103

4.4.3 Location Expectation and Source Detection . . . . . . . . . . . . 108

4.4.4 Spatial Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . 113

4.5 Objective Evaluation of the Proposed BMSS System . . . . . . . . . . . 114

4.5.1 BMSS Empirical Computational Time . . . . . . . . . . . . . . . 114

4.5.2 Experimental Set-up . . . . . . . . . . . . . . . . . . . . . . . . . 116

4.5.3 Objective Performance Measures . . . . . . . . . . . . . . . . . . 117

4.5.4 Detection, Tracking and Separation - BMSS . . . . . . . . . . . . 118

4.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125

5 Conclusions and Further Work 129

5.1 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129

5.2 Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132

5.3 Further Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133

5.3.1 Objective evaluation with controlled speaker movements and speedvariations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133

5.3.2 Dealing with co-locating and crossing over speakers . . . . . . . . 133

5.3.3 Accurate modelling of IVD distribution . . . . . . . . . . . . . . 134

5.3.4 Combining the proposed stationary source separation system withBMSS system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135

A Appdx A: BSS Demonstration 136

A.1 Required Equipment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136

A.2 GUI Layout . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

A.2.1 Device Connected Panel . . . . . . . . . . . . . . . . . . . . . . . 137

A.2.2 Control Panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

A.2.3 Rate of Spatial Filter Adaptation . . . . . . . . . . . . . . . . . . 139

A.2.4 Figures Panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

A.2.5 Sources Panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140

B Intensity Measurement 141

Bibliography 146

Page 9: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures

1.1 Source Separation System . . . . . . . . . . . . . . . . . . . . . . . . . . 2

2.1 Instantaneous Mixture . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.2 Convolutive Mixture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.3 First order microphones showing the sensitivity towards directions ofarrival in the surrounding space (range of sensitivity = 0-1). Figure-of-eight (solid line ζ = 0), Sub cardioid (dashed-dotted line ζ = 0.7),Cardioid (dashed line ζ = 0.5) . . . . . . . . . . . . . . . . . . . . . . . . 26

2.4 First order B-format Audio . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.5 Tetrahedral placement of microphones to form A-format microphone array 28

3.1 Processing stages of the baseline IVD-based system. Initially, the pres-sure and pressure gradient signals are obtained from the microphonearray for intensity measurements. Then, the DFT of the signals is mea-sured. Next, the intensity vector directions (IVD) are calculated in time-frequency domain. Next, the von-Mises spatial filters are applied for theknown source locations. Finally, IDFTs of the separated signals are cal-culated to get the time-domain estimate of separated sources. . . . . . . 34

3.2 The von-Mises spatial filters used for soft masking the time-frequencycomponents based on their IVD measurements. The spatial filters aredepicted with varying circular means and concentration parameters asmentioned in the legend of the figure. . . . . . . . . . . . . . . . . . . . 35

3.3 The A-format microphone array with tetrahedral configuration of thecardioid microphones: Labelled as Left-Forward-Up (LFU), Right-Front-Down (RFD), Left-Back-Down (LBD) and Right-Back-Up (RBU) withrespect to the centre of the configuration. Collectively these are knownas the A-format signals. The distance of the microphones to center ofthe array is denoted by r. . . . . . . . . . . . . . . . . . . . . . . . . . . 36

3.4 The SDR (dB) results of varying angular interval between the sourcesat different beam-widths of the von-Mises spatial filters. The black lineis indicating the trend of increasing SDR as the optimum line. . . . . . . 42

iv

Page 10: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures v

3.5 The SIR (dB) results of varying angular interval between the sources atdifferent beam-widths of the von-Mises spatial filters. The black line isindicating the trend of increasing SIR as the optimum line. . . . . . . . 42

3.6 Average objective evaluation of two source separation by varying spatialfilter beam-width for baseline IVD system. The overall SDR performanceis seen to increase with the increase in spatial filter beam-width (Fig.3.6a). With a greater spatial filter beam-width, more source componentsspread around the desired source location are captured. However, asdepicted in Fig. 3.6b, consequently more components of the interferingsources are captured as well. Due to this a trade-off exists betweensignal to interference ratio and signal to distortion ratio. A suitablebeam-width applicable to all angular intervals is chosen such that SIRdoes not fall below a desired threshold (approx. 20 dB). . . . . . . . . . 43

3.7 Simulation results of location-based space-frequency IVD measurementswith different microphone array sizes. The limit frequency, highlightedusing red vertical line, can be seen to get lower as the array size increases. 46

3.8 Different microphone behaviour depicted for sound originating from dif-ferent locations to visualise the effect of increased non-coincidence basedon source locations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

3.9 Different directivity patterns of the microphone facing towards 45 degree.The sensitivity of the microphone array shown to vary from 0 to 1 fordifferent circular directions of arrival. . . . . . . . . . . . . . . . . . . . . 48

3.10 The recording set-up with distances to floor and the ceiling depicted.The first reflection was found to be approximately 3ms after the arrivaltime of the direct sound component. With the speed of sound to be 340ms it comes out to be around 1m of distance from the closest reflectivesurface, which was the ceiling. . . . . . . . . . . . . . . . . . . . . . . . . 49

3.11 Method to obtain location-based correction coefficients . . . . . . . . . . 50

3.12 Space-frequency microphone response for 70 degrees and 110 degrees.The deteriorating space-frequency response is shown in this figure. Itis represented by the deviation of IVD measurements from the physicallocation of the sources. After a certain limit frequency (approx. 9200 Hzfor the array used) the IVD measurements of both the locations can beseen to overlap in space. As this happens the source content originatingfrom these locations is no longer spatially independent. The response ofthe array at 70 degrees can be seen to phase wrap around after the limitfrequency due to aliasing. . . . . . . . . . . . . . . . . . . . . . . . . . . 52

Page 11: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures vi

3.13 The figure depicts IVD measurements of the microphone array through-out the space. It is the complete space-frequency response, with eachline representing the IVD measurements from a location in space. Againthe space-frequency response is seen to deteriorate significantly after thelimit frequency (approx. 9200 Hz). Four distinguished groups can beidentified as the four quadrants, with space-frequency response converg-ing at 0 degrees, 90 degrees, 180 degrees and 270 degrees. After thelimit frequency is reached (shown as the point of convergence of IVDs)the spatial independence of different locations is compromised due toaliasing at high frequencies. . . . . . . . . . . . . . . . . . . . . . . . . . 53

3.14 Space-frequency response of ITU-R BS1116 standard listening room withT60 of 0.32s at 70 degree and 110 degree locations. The room responseis shown to result in spread of IVD measurements around the physicallocation of the sources. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

3.15 Energy distribution in space of four sources using the microphone arrayresponse only. The energy distribution is observed to be very compactin space. This figure exhibits the potential of IVD-based system withthe microphone array responses low pass filtered at 8000Hz for singlesources located at 40, 130, 220 and 310 degrees. . . . . . . . . . . . . . . 55

3.16 Energy distribution of four sources in ITU-R BS1116 standard listeningroom with T60 of 0.32s. The energy distributions of the sources in spaceare observed to spread around the physical locations of the sources (40,130, 220 and 310 degrees) as a result of the effect of room response. Theenergy distributions without the room response have been depicted inFig. 3.15. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.17 Mean and Standard Deviation of IVD measurements with different re-verberation times - 300Hz to 8000Hz . . . . . . . . . . . . . . . . . . . . 57

3.18 Energy distribution of four simultaneously active sources at 40, 130, 220and 310 degrees, considering only the microphone array responses, hasbeen shown to depict the comparison with energy distributions of singlesources active alone (Fig. 3.15). The effect of overlapping time-frequencycontent among sources results in the spread of energy distributions inspace as compared to compact energy distribution of sources active alone,as shown in Fig. 3.15. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59

3.19 Block diagram of the proposed IVD-based microphone array correctionalgorithm. The IVD measurements are corrected using the look-up tableand source locations before applying the spatial filters. . . . . . . . . . . 60

3.20 Corrected space-frequency response of the microphone array shown inintervals of 10 degrees. The response is observed to improve below thelimit frequency as compared to original IVD measurements depicted inFig. 3.13. See Table 3.2, Table 3.3, Table 3.4 and Table 3.5 for thestatistics of corrected IVD measurements in comparison with the originalmeasurements. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61

Page 12: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures vii

3.21 Mean and Standard Deviation of original (red) and corrected (green)IVD measurements - 300Hz to 8000Hz. See text for discussion. . . . . . 62

3.22 Original and Corrected Mean and Standard Deviation of IVD with dif-ferent reverberation times - 300Hz to 8000Hz . . . . . . . . . . . . . . . 64

3.23 Block diagram of the proposed system to weight the optimum A-formatmicrophones using source locations. . . . . . . . . . . . . . . . . . . . . . 65

3.24 A-format microphone top view. The closest cardioid microphones areidentified for each source location (highlighted in green). Based on theangular interval, weighting of cardioid microphones is proposed to atten-uate unwanted interference and overlapping time-frequency componentsimpinging at the microphone array from opposite direction to that of thedesired source. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

3.25 Block diagram of the proposed system. The Energy and IVD measure-ments are selected based on their closeness to desired source locationusing regions of interest. Once the local statistics of the source are lo-cally isolated, spatial filters parameters are set accordingly. . . . . . . . 68

3.26 Regions of interest for four source locations. The region of interest isan ideal circular binary filter to locally isolate the IVD measurements ofdesired sources. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69

3.27 Objective evaluation: Adaptation rate of spatial filter parameters . . . . 73

3.28 Block diagram of the proposed IVD-based source separation system (con-tributions highlighted in green). Initially, the pressure and pressure gra-dient signals are obtained from the microphone array. Then, the DFT ofthe signals are calculated. Next, the intensity vector directions are cal-culated and consequently the IVD measurements are corrected using themicrophone array look-up table. Next, energy weighted mean and stan-dard deviation are used to adapt spatial filter parameters in regions ofinterest that locally isolate IVD measurements of each source. Based onknown source locations, the cardioid weight matrix is determined. Forthe desired sources, adapted von-Mises spatial filters are used to per-form separation using appropriately weighed microphone signals basedon cardioid weight matrix. Finally, the IDFTs of the separated signalsare calculated to get time-domain estimates of the sources. . . . . . . . 75

3.29 Computational time of different processing steps in the proposed systempipeline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

3.30 SDR comparison baseline IVD system performance (red) with (a) Micro-phone array correction (b) Adaptive spatial filters (c) Cardioid weightmatrix (d) Combining all modifications, for two source scenario. . . . . . 78

3.31 SIR Comparison of baseline IVD system performance (red) with (a) Mi-crophone array correction (b) Adaptive spatial filters (c) Cardioid weightmatrix (d) Combining all modifications, for two source scenario. . . . . . 80

Page 13: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures viii

3.32 Comparison of baseline IVD system performance (red) with all modifi-cations combined (blue), for two source scenario, RT = 0.67s . . . . . . 81

3.33 Different four sources scenarios considered for experiments - Minimumangle between any two speakers varying from 30 degree to 90 degreein 20 degree intervals to check systems performance for varying angularintervals. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

3.34 SDR Comparison of baseline IVD system performance (red) with (a)Microphone array correction (b) Adaptive spatial filters (c) Cardioidweight matrix (d) Combining all modifications, for four source scenario. 83

3.35 SIR Comparison of empirically best selected baseline performance (red)(a) Microphone array correction (b) Adaptive spatial filters (c) Cardioidweight matrix (d) Combining all modifications, for four source scenario. 84

3.36 Comparison of baseline IVD system performance (red) with all modifi-cations combined (blue), for four source scenario, RT = 0.67s . . . . . . 85

4.1 First order IIR filter structure . . . . . . . . . . . . . . . . . . . . . . . . 92

4.2 IVD Distribution in space (a) Without von-Mises modelling (b) Apply-ing first order IIR Filter to smooth out noise in the time domain (c)Modelling the IVD with von-Mises distribution funtion to smooth outnoise in space (d) Applying von-Mises model along with first order IIRfilter to achieve denoised IVD distribution in both time and space. . . . 93

4.3 Effect of distance of the speaker to microphone array on IVD distributionin space. (a) Ground truth (b) IVD distribution in space. Speaker 1 ismoving forwards and backwards w.r.t to the microphone array whileSpeaker 2 is stationary in space. The closer the speaker gets, the greaterits captured energy and its direct to reverberant ratio. As a result ithas better IVD distribution representation in space (prominent with ahigh peak). The distribution of a physically stationary Speaker 2 is alsoshown to get affected due to the change in distance of Speaker 1. . . . . 94

4.4 IVD distribution of two moving speakers at the same distance to the mi-crohone array. (a) Ground truth (b) IVD distribution in space. Speaker1 is moving between 75 degrees and 125 degrees, while Speaker 2 ismoving between 175 and 225 degrees. . . . . . . . . . . . . . . . . . . . 96

4.5 Processing stages of the baseline IVD-based system (see Section 3.2).Initially, the pressure and pressure gradient signals are obtained fromthe microphone array. Then, the DFT of the signals is measured. Next,the intensity vector directions are calculated in time-frequency domain.Next, von-Mises spatial filters are applied for the known speaker loca-tions. Finally, IDFTs of the separated signals are calculated to get thetime-domain estimates of the target speakers. . . . . . . . . . . . . . . . 97

Page 14: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures ix

4.6 The utilisation of time-frequency units up to 4000Hz to up to 22050Hzfor IVD distribution representation. Speaker 2, which is distant (approx.3 meters) to the microphone array portrays less prominent IVD distri-bution with increase in the frequency range until it completely becomespart of the IVD distribution floor. Also with the increase in frequencyrange, IVD distribution of Speaker 1 can be seen to spread in space. . . 99

4.7 Log IVD distribution with IIR filtering corresponding to different aver-aging times (a) 46.4 ms (b) 185.8 ms (c) 278.6 ms (d) 371.5 ms. As theeffective averaging time increases, the IVD distribution becomes moreregular and denoised to reveal reliable estimates of the speakers; how-ever, averaging over a much longer period of time leads to poor speakerestimates, because longer averaging times may result in multiple peaksfor the same speaker due to a slow update rate of the IVD distribution(4.8). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101

4.8 Peak Detection & Location Expectation with IIR filtering correspondingto different averaging times (a) 46.4 ms (b) 185.8 ms (c) 278.6 ms (d)371.5 ms. The red cross indicates the time delay caused in peak detectiondue to longer averaging times. See text for discussion. . . . . . . . . . . 102

4.9 Average IVD distribution measurements over 2.5s of recorded speechactivity: (a) Two Speakers (b) Four Speakers. The average IVD distri-bution peaks are higher and more prominent for less speakers. With agreater number of speakers the IVD noise floor gets high due to addedreverberations and increase in overlapping time-frequency content ofspeakers, resulting in a poor IVD distribution representation. Due topoor IVD representation over short time durations, added speakers posea challenge to the moving blind source separation system. . . . . . . . . 104

4.10 (a)-(c) IVD distribution for different number of speakers (d)-(f) PeakDetection results - high static threshold = 1.5 (g)-(i) Peak Detectionresults - adaptive to number of speakers detected (j)-(l) Peak Detectionresults - low static threshold = 0.5. (Red ellipses show errors. SeeSection 4.4.2.1 for discussion.) . . . . . . . . . . . . . . . . . . . . . . . . 107

4.11 Peak detection and location expectation using different increment valuesin the location expectation using p. The false peaks are circled in red forclarity in analysis. The amount of false detections can be observed toreduce as the delay in speaker detection increases from an instantaneousvalue to 0.3s. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

4.12 Peak detection and location expectation using different increment valuesin location expectation using p. Red arrows represent delay in time whiledetecting a new speaker. See the text for discussion. . . . . . . . . . . . 111

4.13 Peak detection and location expectation using different decrement valuesin the location expectation by varying q. Green arrows show the waitingtime before declaring a speaker as absent. See the text for discussion. . 112

Page 15: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures x

4.14 Block Diagram of Proposed Moving BSS System (contributions high-lighted in blue). Initially, the pressure and pressure gradient signals areobtained from the array. Then, the DFT of the signals are calculated.Next, the intensity vector directions are calculated and consequentlyIIR filtered IVD distribution in space is measured using von-Mises mod-elling of the IVD measurements. Next, peak detection is done to locatepotential speakers in the surrounding space. Location expectations areassigned to the identified peaks for reliable speaker detection and to caterfor pauses in between speech activity. For the detected speakers time-variant von-Mises spatial filters are designed and applied. Finally, IDFTof the separated signals are calculated to get the time-domain estimatesof target speakers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113

4.15 Computational time of proposed BMSS algorithms with respect to num-ber of speakers. See the text for discussion. . . . . . . . . . . . . . . . . 115

4.16 Experimental Setup: (a) Stationary source locations (b) Regions ofspeaker movement. A maximum 60 degree interval of speaker move-ment is assigned for each region in the designed set-up. It is to be notedthat the proposed system can perform speaker detection, tracking andseparation even if the speaker moves out of the allocated region, as longas it does not co-locate with other active speakers. . . . . . . . . . . . . 116

4.17 Examples of two moving speakers - Tracking and detection performedby the proposed BMSS system. In (a) and (c) the location expectationof speakers is depicted to show tracking. Separate ‘tracks’ of speakermovement can be observed in space. Near the time frames in the middle(800 - 1200), the speakers take a long pause and start speaking againwhile carrying out movements. In (b) and (d) the true speaker activityis compared with the activity determined by the proposed BMSS system. 119

4.18 Two speaker mixtures: PEASS Toolbox results showing comparisonbetween baseline IVD implementation and BMSS system with movingspeakers and stationary speakers. See text for discussion. . . . . . . . . 122

4.19 Two speaker mixtures: BSS Eval Toolbox results showing comparisonbetween baseline IVD implementation and BMSS system with movingspeakers and stationary speakers. See text for discussion. . . . . . . . . 123

4.20 Examples of four moving speakers - Tracking and detection performed bythe proposed BMSS system. In (a) the location expectation of speakersis depicted to show tracking. In (a) separate detected ‘tracks’ of thefour speakers in space can be observed. Near the time frames in themiddle (800 - 1200), the speakers take a long pause and start speakingagain while carrying out movements. In (b) the true speaker activitiesare compared with those determined by the proposed BMSS system. . . 124

4.21 Four speaker mixtures: PEASS Toolbox results showing comparisonbetween baseline IVD implementation and BMSS system with movingspeakers and stationary speakers. See text for discussion. . . . . . . . . 126

Page 16: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures xi

4.22 Four speaker mixtures: BSS Eval Toolbox results showing comparisonbetween baseline IVD implementation and BMSS system with movingspeakers and stationary speakers. See text for discussion. . . . . . . . . 127

A.1 Graphical user interface of BSS application . . . . . . . . . . . . . . . . 137

A.2 Device connected panel . . . . . . . . . . . . . . . . . . . . . . . . . . . 138

A.3 Control Panel in GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138

A.4 Rate of spatial filter adaptation panel . . . . . . . . . . . . . . . . . . . 139

A.5 Figures panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

A.6 Sources panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140

Page 17: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Nomenclature

Symbols

c (t) Threshold for peak detection

K Average approximated IVD distribution floor

rn Resultant vector for source n

D Diagonal scaling matrix

I Permutation matrix

δi (f, t) Relative delay between microphones

εn Location expectation of speaker n

d Smallest distance considering Nyquist theorem

F A-format to B-format conversion matrix

f (θ) Microphone response at angle θ

I Identity matrix

k Window sample index

M Window size in samples

m Mask used in time-frequency filtering

P Power

sn Separated source n

W Additive noise data

w Window function

Λ Cross-section area

λmin Smallest wavelength considering Nyquist theorem

µ Mean circular direction of von-Mises distribution

xii

Page 18: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures xiii

ν Volume

Øn Region of Interest for source n

ω Angular Frequency

φ Phase

ρ Density

σ Concentration of von-Mises distribution around mean location

θ Circular variable for von-Mises distribution

θd Angular interval of source location from cardioid microphone orien-tations

θBW Beam-width value

θLBD LBD cardioid microphone orientation angle

θLFU LFU cardioid microphone orientation angle

θRBU RBU cardioid microphone orientation angle

θRFD RFD cardioid microphone orientation angle

ζ Microphone weighting to determine directional response

A Extent of displacement

a Acceleration

Af A-format data matrix

ai (f, t) Relative amplitude between microphones

B Unmixing matrix

c Speed of sound

Cn(ω, t) Corrected time-frequency IVD for source n

D Energy Density

d Array vector

E Energy

Ek Kinetic Energy

Ep Potential Energy

f (θ;µn, σn) Spatial filter with µ direction and σ beam-width for source n

F Force

Page 19: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures xiv

f Frequency

Gi Cluster centre representing source i

H Room Transfer Function Data

h Elastic Constant

I Intensity

Jn Subset of Cn based on region of interest

k Wave number

Ln Index of correction vector for source n

M Number of microphones

m mass

N Number of sources

P Pressure

p Increment value in speaker location expectation

p(si) Probability density function of i-th source

p (s1, s2, s3...sN ) Joint probability density function of sources

PBLD A-format Back-Left-Down pressure signal

PBRU A-format Back-Right-Up pressure signal

PFRD A-format Front-Right-Down pressure signal

PLFU A-format Left-Front-Up pressure signal

Q Resolution factor of correction

q Decrement value in speaker location expectation

r Radial Distance

r (θ, f) Beam-former response at angle θ and angular frequency ω

S Source data

t Time

U Cardioid weight matrix

V Velocity

W Inverse of room transfer function

w Unmixing weight vector

Page 20: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures xv

WB B-format omnidirectional response

X Mixture Data

XB B-format figure-8 response across x-axis

Y Separated source data

YB B-format figure-8 response across y-axis

Zn(ω, t) IVD correction vector for source n

Abbreviations

ASIO Audio sound input output

BSS Blind Source Separation

CASA Computational Auditory Scene Analysis

CLT Central Limit Theorem

DFT Discrete Fourier Transform

DOA Direction of Arrival

DUET Degenerate Unmixing Estimation Technique

ECG Electrocardiogram

EEG Electroencephalography

GPS Global Positioning System

GUI Graphical user interface

ICA Independent Component Analysis

IDFT Inverse discrete Fourier transform

ILD Inter-aural level difference

ITD Inter-aural time difference

IV D Intensity Vector Direction

ML Maximum Likelihood

MMSE Minimum Mean Square Error

OS Operating system

PCA Principle Component Analysis

PDF Probability Density Function

RIR Room Impulse Response

Page 21: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

List of Figures xvi

SI Standard International

SIR Signal to Interference Ratio

STFT Short Time Fourier Transform

WNG White Noise Gain

Page 22: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Chapter 1

Introduction

Audio information is processed during many activities in the daily life of a human.

Through perceived acoustic signals, people are able to extract valuable information

about their surrounding environment. Even in a reasonably crowded room, a person

with healthy hearing is able to identify and focus on a particular acoustic source of

interest. Such source of interest may be a musical instrument, a single speaker or

multiple speakers having a conversation among a number of interfering sources. The

ability of humans to perform such acoustic tasks is referred to as the solution to the

cocktail party problem [1].

Intense research has been conducted in the domain of digital signal processing over last

few decades to try and mimic the human ability to perform a range of acoustic tasks.

These include source detection, source identification, source localisation, source tracking

and source separation. A result of such research is the Computational Auditory Scene

Analysis (CASA). CASA is the approximation of human auditory system to localise

and isolate acoustic sources of interest [2]. Other approaches also exist which try to

solve the problem of source separation purely based on mathematics. These approaches

try to exploit source signal properties and the environment transfer function.

A significant number of source separation techniques rely on the microphone arrays to

perform separation [3][4]. Such arrays comprise of a number of microphones arranged

in a single or multidimensional configuration. This thesis deals with a compact coinci-

dent microphone array, comprised of four cardioid microphones placed in a tetrahedral

1

Page 23: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2

Source Separation System

Source 1

Source N

Source 1

Source N

Figure 1.1: Source Separation System

configuration. Mixture signals recorded by the microphone array are processed with

source separation algorithm to output individual sources in separate channels.

Blind Source Separation (BSS) refers to the problem of source separation when no

prior mixing information is available. This information can be in the form of number

of sources, the location of sources, knowledge of Room Impulse Response (RIR) and/or

any target source signal information. In more challenging source separation scenarios,

added information includes the speed with which sources and/or the microphone array

moves, and the time at which a particular source enters and/or leaves the acoustic

scene. In a real environment, the source signals are not only directly captured by

the microphone array, but the time delayed reflections are also part of the recordings.

This further complicates the source separation problem. The information of mixing

is contained in the RIR [5]. Most commonly, the RIR contains direct sound from

sources and, their reflections and reverberations as a result of source interaction with

the surrounding environment.

Some of the source separation techniques utilise spatial diversity to perform separation

of sources [6]. Having the knowledge of target source location or interfering sources, a

desired spatial response can be constructed to extract the source of interest. Adaptive

methods have also been proposed which modify the spatial response of the microphone

array according to the perceived source signals [3][4]. A different approach based on

Page 24: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

1.1. Applications 3

the statistical properties of microphone signals rather than spatial diversity also exists

[7][8]. These different types of algorithms are still widely under investigation. The

choice of algorithm is purely based on the application requirement and implementation

environment. The requirements include, but are not limited to, real-time or off-line

processing [9][10], compactness of the microphone array [11] and the number of sources

to be separated [12][13]. For example, in a small electronic device such as a hearing

aid, the integration of a compact microphone array is desired.

Blind source separation is a multifaceted problem, as such, the issues that need to

be addressed include degraded performance with increase in reverberation time [14],

sources located close to one another [11], coping with greater number of sources [15],

localisation of sources [16], source activity detection and most importantly the sce-

nario of moving sources [17][18]. Over the past decade considerable research has been

carried out for physically stationary sources [19][13][20][21][22]; comparatively, the par-

ticular problem of separation of moving sources has not been extensively investigated

[23][18][24]. It is unrealistic to consider a source as stationary in daily life scenarios,

for example, a professor may move around while delivering a lecture. The movement of

sources poses a challenging problem to the researchers. The aim is to constantly detect

changes in mixing conditions and update unmixing system parameters accordingly with

a low latency, to maintain output sound quality of the source of interest.

1.1 Applications

Blind source separation is essentially a complex mathematical problem. The solution

of this problem has direct impact on many applications of signal processing. It can

be used in video conferencing, to obtain desired speakers voice and possibly direct the

camera towards the speaker of interest using localisation cues.

Speech and speaker recognition systems need interference free voice to perform effective

recognition. In a car while giving voice commands, the cell phone or a Global Position-

ing System (GPS) does not necessarily capture only drivers voice. It is possible that

interfering sources are active, or the noise of car engine or a nearby air-conditioning

system may corrupt the drivers voice beyond recognition. As such, in hands-free human

Page 25: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

1.2. Scope and Objective 4

computer interaction, it is essential to separate the speaker of interest from unwanted

interfering sources and carry out voice commands. Similar concept can be applied to

suppress interfering sources while carrying out communication over the internet, for

example while communicating over Google Hangout or Microsoft Skype.

The ability of humans to listen and process sound present in the surrounding environ-

ment is extremely important in day to day life. Some human beings lose the ability to

hear as they get older. Not only low energy sounds become hard to recognise, frequency

dependent loss may also develop. When multiple speakers are active at the same time,

it becomes more difficult for humans with less ability to hear out the speaker of interest.

High quality hearing aids with blind source separation algorithm are used to reduce

their listening effort and improve speech intelligibility.

Acoustic surveillance is another important application of blind source separation. It

can be used by security agencies for spying and intelligence related purposes. The

separated output and source localisation cues can be used in forensic analysis as well.

Also, in underwater acoustics blind source separation has the ability to detect, separate

and facilitate understanding of underwater acoustic events. These include tracking ship

movements, and detecting gas or oil leakages.

1.2 Scope and Objective

Aiming to achieve a deterministic and low complexity source separation system, the

Intensity Vector Direction (IVD) based algorithm was developed by Gunel et al. [11].

The purpose of this thesis is to further the field of source separation by investigating

and improving the IVD-based source separation system for practical acoustic scenarios.

As such, this thesis deals with analysis and improvement in applicability and output

sound quality of the IVD-based system, while dealing with stationary and moving

sources scenarios.

The scope and objectives of this thesis include:

1. To fully understand and investigate the existing IVD-based source separation

system.

Page 26: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

1.3. Thesis Outline 5

2. To propose new algorithms for improving output sound quality and applicabil-

ity of existing IVD-based source separation system, wherever shortcomings are

identified.

3. To investigate existing source separation system to perform blind source separa-

tion in practical acoustic scenarios, including but not limited to moving sources.

4. To propose new algorithms to develop IVD-based source separation system in

order to perform separation in practical acoustic scenarios, including but not

limited to moving sources.

1.3 Thesis Outline

Over the past few decades several types of source separation techniques have been

proposed which deal with physically stationary sources, while only a few have been

proposed to deal with moving sources. Related techniques in the field of source separa-

tion are discussed in Chapter 2. Source separation techniques can be widely categorized

into three distinct groups. These are the statistical, beam-forming and time-frequency

sparsity-based source separation techniques, presented in Section 2.1.2, Section 2.1.3

and Section 2.1.4 respectively. The background of sparsity-based IVD technique is

highlighted in Section 2.2 to facilitate the understanding of its foundations.

Analysis and improvement of IVD-based source separation system for stationary sources

scenario is presented in Chapter 3. The baseline IVD-based source separation technique

is presented and evaluated in Section 3.2 to support the novelty of contribution chap-

ters. The system is investigated with respect to microphone array (Section 3.3.2),

surrounding room environment (Section 3.3.3) and impact of multiple sources on the

time-frequency IVD measurements (Section 3.3.4). Algorithms are proposed to com-

pensate for the non-ideal behaviour of microphone array (Section 3.4), select the most

suitable microphones to separate the source of interest with a superior output qual-

ity (Section 3.5) and adaptively cater for the room response and interfering sources

(Section 3.6).

Page 27: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

1.4. Original Contributions 6

Later, the IVD-based source separation system is investigated to perform in practi-

cal acoustic scenarios, which include time-variant activity of the sources and moving

sources. In Section 4.3, a new analysis with respect to IVD distribution in space is

presented for practical acoustic scenarios. The analysis is carried out to recognise and

exploit the full potential of existing IVD-based system in order to perform source activ-

ity detection and tracking. To develop IVD-based system into a blind source separation

system capable of dealing with practical acoustic scenarios, denoised IVD distribution

(Section 4.4.1), adaptive peak detection (Section 4.4.2), and location expectation (Sec-

tion 4.4.3) algorithms are proposed.

Finally, the IVD-based blind source separation demonstration developed in Matlab

2012b with the proposed algorithms is presented in Appendix A. The graphical user

interface (GUI) of the developed application is discussed in Appendix A.2. It facilitates

setting up the recording equipment and provides GUI to interact with surrounding

acoustic sources in real-time. The detected and tracked sources are displayed as options

on GUI to allow the user to choose among several active sources (Appendix A.2.5). As

a result the sources can be listened to in any desired combination and real-time plots

on the GUI show the IVD statistics as they vary with time (Appendix A.2.4).

1.4 Original Contributions

1. A new analysis of the IVD-based source separation system is presented utilising

a real microphone array measurements and room impulse responses.

2. A new measurement-based microphone correction algorithm is proposed in Sec-

tion 3.4 to provide better localisation cues, and it is evaluated for two different

room environments.

3. A new algorithm is proposed to select and weight cardioid microphones based on

the desired source locations (Section 3.5). It is proven to give superior output

sound quality.

4. An energy weighted algorithm is proposed to adaptively update the spatial filter

parameters in a non-exhaustive manner, based on local statistics of the IVD

Page 28: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

1.5. Publications by the Author 7

measurements (Section 3.6).

5. A new analysis to deal with practical acoustic scenarios with respect to IVD

distribution in space is presented in Section 4.3.

6. Based on IVD statistics, a new blind moving source separation system pipeline,

capable of speaker detection, localisation, tracking and eventually separation, is

proposed to deal with practical acoustic scenarios in Section 4.4.

1.5 Publications by the Author

1. A. Riaz, X. Shi, A. Kondoz, “Adaptive Blind Moving Source Separation based

on Intensity Vector Statistics”, submitted to IEEE/ACM Transactions on Audio,

Speech and Language Processing.

2. A. Riaz, X. Shi, A. Kondoz, “Adaptive Source Separation of Convolutive Mixtures

based on Intensity Vector Statistics”, to be submitted, IEEE/ACM Transactions

on Audio, Speech and Language Processing.

Page 29: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Chapter 2

Related Work

Acoustic blind source separation is the problem of automated separation of audio

sources from observed mixtures. It is commonly known as the cocktail party prob-

lem. The cocktail party problem is described as focusing on one active speaker in a

room where multiple speakers are simultaneously active. It was first clearly defined

by Colin Cherry [25]. For humans it is not difficult to pay attention to one of the

speakers; however, for machines the problem is still being widely investigated. Re-

search has been carried out over the past few decades to study several aspects related

to the cocktail party problem. These include localisation of sound sources, geometry of

the microphone array, processing complexity of the algorithm, room impulse response

identification, identifying time-variant mixing system parameters for moving sources,

approximating time-variant number of sources and estimating any target source signal

information. Although this thesis focuses on acoustic BSS, the mathematical implica-

tion of these techniques is useful in other areas as well, such as in image processing and

medical signal processing [26].

In this chapter, first the related work in the field of source separation is presented. It

is followed by the background of IVD-based source separation system and a discussion

on the microphone array used for intensity measurement.

8

Page 30: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 9

2.1 Related Work

The most widely used methods for performing source separation are the statistical

methods, such as ICA [7][8]. Second category of methods rely on the microphone array

configuration entirely, for example the adaptive beam-forming [3][4]. The remaining

methods, like DUET [27], exploit time-frequency sparsity among source signals to per-

form separation.

The BSS problem for M microphones and N sources for stationary sources scenario

can be generally represented as [28]:

X(t) = HS(t), (2.1)

where N number of source signals are S(t) = [s1(t)s2(t)s3(t)...sN (t)]T . The trans-

fer function coefficients are contained in (M × N) mixing matrix H, such that hmn

represents coefficient for m-th microphone and n-th source. X(t) = [x1(t)x2(t)x3(t)...

xM (t)]T contains the resultant mixture components of M microphones. The task is

to extract source signals S(t) from observed X(t) mixtures. The two mixture models

which are considered while solving the blind source separation problem are presented

first. It is followed by a discussion on blind source separation techniques that are being

actively investigated.

2.1.1 Source Mixture Models

The instantaneous and convolutive mixture models define the mixture signals encoun-

tered while dealing with source separation problem.

2.1.1.1 Instantaneous Mixture Model

The instantaneous mixture model assumes that the sensors measure only scaled version

of the sound sources at every time instant t. For three sources and three microphones

scenario, this model is mathematically represented as [29]:

Page 31: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 10

x1(t)

x2(t)

x3(t)

=

h11 h12 h13

h21 h22 h23

h31 h32 h33

s1(t)

s2(t)

s3(t)

+

w1(t)

w2(t)

w3(t)

. (2.2)

The sources are assumed to be physically stationary in this representation. s1(t), s2(t)

and s3(t) represent the source signals and wn(t) represents the additive noise. h12 rep-

resents the time-invariant transfer function coefficient from source s1 to the microphone

x2. Likewise, other transfer function coefficients between sources and microphones are

contained within the hmn coefficients matrix. x1(t), x2(t) and x3(t) are the resultant

acquired mixtures. For a general scenario with M microphones and N sources, the

representation for instantaneous mixture model can be done as:

X(t) = HS(t) + W (t). (2.3)

The instantaneous mixture model is the simplest form of modelling a mixture of signals.

It implies that source signals are only being amplitude modulated, not time delayed

or echoed. It represents an ideal surrounding environment where reverberation or

multipath propagation is absent. The instantaneous mixture model is most commonly

used for research purposes while dealing with acoustic applications. Several other

signal processing applications of the instantaneous mixture model also exist [30][26],

such as in image processing, to extract independent features of an image in order to

denoise and improve overall image quality [26]. It is also applicable in medical signal

processing, for example when the underlying components of brain activity have to be

identified from given recordings in the form of an electroencephalogram (EEG) [30]. The

ideal instantaneous mixture model with respect to acoustic sources and microphones is

depicted in Fig. 2.1.

2.1.1.2 Convolutive Mixture Model

In practical acoustic scenarios, source separation algorithms must take convolutive

mixing of the acoustic sources into account, as the multipath effects of real reverberant

Page 32: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 11

Figure 2.1: Instantaneous Mixture

environment cannot be ignored [31]. As compared to instantaneous mixture model, the

received signals not only contain scaled version of source signals but also their filtered

and dispersed versions. In a reverberant room, this phenomenon takes place due to

different times of arrival of direct sound, reflected sound and reverberating sound. It

makes source separation problem much more challenging to solve, especially when the

reverberation time is great. The mathematical representation of three sources and three

microphones scenario for a convolutive mixture model is [29]:

x1 (t) = [h11 (t)⊗ s1 (t) + h12 (t)⊗ s2 (t) + h13 (t)⊗ s3 (t)] + w1(t) (2.4)

x2 (t) = [h21 (t)⊗ s1 (t) + h22 (t)⊗ s2 (t) + h23 (t)⊗ s3 (t)] + w2(t) (2.5)

x3 (t) = [h31 (t)⊗ s1 (t) + h32 (t)⊗ s2 (t) + h33 (t)⊗ s3 (t)] + w3(t). (2.6)

Here ⊗ denotes convolution, sn (t) represents source signals, hmn (t) represents the

transfer function between the n-th source and m-th microphone, and xm(t) represents

the obtained mixtures. wm(t) represents the additive noise component. Using matrix

notation it can be written as:

X (t) = [H (t)⊗ S (t)] + W (t). (2.7)

Convolution in time domain corresponds to multiplication in the frequency domain,

therefore equation (2.7) can be rewritten for frequency domain representation as:

X (ω) = [H (ω)S (ω)] + W (ω). (2.8)

The scenario of convolutive mixtures is depicted in Fig. 2.2.

Page 33: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 12

Figure 2.2: Convolutive Mixture

2.1.2 Statistical Methods

Independent Component Analysis (ICA) was first clearly defined in 1994 by P. Comon

[8] after initial work done with Jutten and Herrault [32]. It is a statistical technique,

used to expose underlying hidden components from sets of observed measurements.

ICA can be seen as an advancement of Principal Component Analysis (PCA). In its

simplest form it was proposed to solve the instantaneous mixture model; however,

for real environments ICA has to deal with convolutive mixtures and it increases in

computational cost due to increase in FIR filter length. To overcome this problem,

frequency domain implementation of the ICA algorithm is carried out [31]. ICA for

instantaneous mixture model forms the basis for dealing with convolutive mixtures. It

is essentially parallel execution of a series of instantaneous ICA at each frequency bin.

In order for ICA to perform unmixing, some assumptions need to be made [33]. Firstly,

the source components are assumed to be statistically independent. Mathematically it

implies that the joint Probability Density Function (PDF) is factorizable as [34]:

p (s1, s2, s3...sN ) =N∏i=1

p(si). (2.9)

For example two random variables, sx and sy, are said to be independent if there is no

mutual information among the two, i.e. information of sx does not give any information

of sy, and vice versa. Although the source signals are independent, the mixture signals,

xm, are not. As such, the underlying independent source components are extracted by

Page 34: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 13

applying ICA on mixture signals. ICA performs source separation by exploiting higher

order statistics of the signals. The higher order cumulants of the Gaussian distribution

are zero, implying that independent components of each source should ideally have

non-Gaussian distributions.

Another assumption made for ICA is that the mixing matrix is invertible. As a result,

the number of sources is equal to or less than the number of microphones. For specific

scenarios, where the number of microphones and sources may vary, M greater than N

case is termed as overdetermined, M less than N is termed as underdetermined and

M equal to N is termed as a determined case. The time-frequency content of sources

can be exploited to solve the underdetermined case as well [13]; however, it requires

additional processing steps.

Considering the noise free version of equation (2.3), the ICA solution exists in having

an unmixing matrix W such that:

S(t) = WX(t). (2.10)

The time relation will be dropped in the following representations for simplicity. Coef-

ficients of W can be found using prior measurements of the transfer function H. This

however is not a practical solution. ICA instead finds the coefficients of B such that:

BH = ID, (2.11)

where D is a diagonal scaling matrix and I is an (N ×N) permutation matrix. Hence,

the separated statistically independent sources can be found as:

Y = BX. (2.12)

Here Y represents the statistically independent components of separated sources, ob-

tained from processing the mixture signals, X. There are several techniques for solving

the instantaneous ICA problem.

Page 35: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 14

2.1.2.1 PCA Preprocessing

For performing ICA, the sources are assumed to be independent, and as such, their

joint PDF can be represented in the form of equation (2.9). PCA preprocessing can be

used to decorrelate the mixtures X prior to performing ICA [8]. Independence implies

decorrelation and not the vice versa. PCA preprocessing facilitates ICA by performing

decorrelation over the lower order statistics. It results in a transformation of X such

that:

E{xxT

}= I , (2.13)

where x represents the transformation of X and it is whitened data with resulting

correlation matrix I, a unit matrix. This is done only for dealing with the second order

statistics of the signals. In order to achieve complete independence, the signals are

to be processed for all orders using ICA. Therefore PCA is used as a preprocessing

stage for ICA to decorrelate the data, reduce noise, reduce computations and achieve

a quicker convergence [35].

2.1.2.2 ICA-based Techniques

ICA has developed as a general concept representing a family of techniques which

aim to extract independent source signals. From the central limit theorem (CLT) it

is known that the sum of independent signals tend to be more Gaussian than any of

the two signals considered alone [36]. Hence independence implies non-Gaussianity, as

such, in ICA an unmixing matrix (W ) is found such that non-Guassianity of individual

source signals is maximised. ICA removes dependence in higher orders (after PCA

preprocessing) and hence ICA is often termed as the rotation which makes decorrelated

data independent [26]. For achieving independence, higher order statistics such as

kurtosis are used. Iterative methods are utilised to either maximise or minimise the

considered cost function, such as related to kurtosis [37] or mutual information entropy

[8]. All the cost functions relate to non-Gaussianity to an extent. As such, ICA has

been formulated with several implementation approaches:

Page 36: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 15

1. Maximum Likelihood (ML) [38][39][40]

2. Maximisation of information transfer [41][42][43]

3. Negentropy [44]

4. Higher order moments and cumulants [8][45]

5. Non-linear PCA [46][47][48]

These approaches are based on cost functions that try to exploit non-Gauassanity of

source signals. The reasons for using different cost functions is related to computational

cost or sensitivity to outliers. There are several gradient descent methods available for

using these cost functions in order to derive unmixing matrix, W . These include Newton

method of descent [35], steepest gradient descent and stochastic gradient descent [49].

The natural gradient descent method is one of the most commonly used methods for

performing ICA [50].

ICA algorithms for instantaneous mixture model can be applied at each frequency bin in

frequency domain to deal with convolutive mixtures. Commonly, a Short Time Fourier

Transform (STFT) is used for framing the time data and transforming each time frame

into the frequency domain. As such, the representation of signals is carried out as

x (f, t), h (f, t) and s (f, t) exhibiting a dependence on both the time and frequency

[51].

2.1.2.3 Limitations of ICA

In ICA there are ambiguities with regards to order and scaling of the separated source

signal components. It is due to the permutation matrix, as given in equation (2.11). The

problem extends to each frequency bin while dealing with convolutive mixtures. The

permutation ambiguity is described as the mixing up of estimated independent sound

source components due to lack of proper alignment [13]. The permutation ambiguity is

faced because both S and H (equation 2.1) are unknown, so the order of components

can get exchanged even after the independent source components have been determined

[52].

Page 37: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 16

Although many post processing algorithms have been proposed to designate the sep-

arated independent components to their corresponding sources, permutation problem

has no closed form solution [53]. The proposed solutions either utilise time-frequency

source models, exploit room impulse response or use the array geometry information

[31][54][55][56][57][52]. The post processing required to solve permutation problem adds

to the computational cost, and as no closed form solution exists, it is still being widely

investigated [58][53].

Another ambiguity faced by ICA is the scaling ambiguity [52]. Although it is not

as severe as the permutation ambiguity, it occurs when the energy of independent

components does not relate to the energy of source signals. Since both S and H are

unknown, any scaling of source signal Si could be made ineffective by scaling down the

corresponding column Hi in H:

X =∑i

(1

αHi

)(αSi) . (2.14)

This problem is commonly resolved by normalising the unmixing matrix W [59].

Another issue with ICA is the assumption of linearly stationary mixing conditions, i.e.

the mixing matrix is assumed to be constant, independent of the time-variant source

signals [60]. This assumption degrades the effectiveness of ICA in real life scenarios, for

example, during electrocardiogram (ECG) the mixing conditions change over time when

the person inhales or exhales [61]. Also, in acoustic applications, moving sources lead

to changes in the mixing conditions [62], which are inherently assumed to be stationary

by ICA. As such, this assumption creates problems while using ICA in more practical

scenarios.

At the final stage, the approximated unmixing matrix W is used to perform separation

in the time-frequency domain using STFT as:

y (f, l) = W (f, l)x (f, l) , (2.15)

where y (f, l) = [y1 (f, l) , y2 (f, l) ...yN (f, l)]T contains the source estimates, W (f, l)

is time-frequency representation of unmixing matrix and x (f, l) = [x1 (f, l) , x2 (f, l) ...

xM (f, l)]T represents the obtained sound mixtures.

Page 38: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 17

2.1.3 Beam-forming Methods

Despite using microphone array information for solving the permutation problem, ICA

methods fundamentally rely on statistical independence of source signals [8]. A separate

category of algorithms exists which performs source separation based on the knowledge

of microphone array geometry [3][4]. Commonly termed as beam-forming [63], these

methods can be used to spatially separate out data from the desired locations. The

sensors of the array are located at different points and as a result are able to sample

the incoming sound wave in space. The acquired data is then processed to attenuate

interfering signals and extract desired source signal. Some of these techniques focus on

finding the Direction Of Arrival (DOA) to perform spatial filtering.

The beam-forming methods work on the same principle as that of a simple digital filter.

In a digital filter the signals are passed or stopped depending on their frequency content.

Spatial diversity can be exploited in a similar manner of passing or stopping the source

signals depending on the direction of arrival. It is commonly achieved by delaying parts

of the input signal and multiplying by weights. In a beam-former, simple tap delays

correspond to the weights that result in spatial filtering. Despite having overlapping

frequency content, the signals can be recovered due to their different locations in the

surrounding environment [63].

Using beam-forming technique for the desired source, a weight vector w is used in the

frequency domain to perform source separation [63]:

y (f) = w (f)H x (f) , (2.16)

where x (f) is the mixture and y (f) is the separated source. Here H denotes the

Hermitian operator. Due to geometry of the microphone array, there is a physical

delay in time of arrival of sound at each microphone, given by [63]:

d (θ, f) =[1, e−j2πfτ1 , e−j2πfτ2 , ... , e−j2πfτM−1

]T. (2.17)

The physical delay is a function of the source direction θ and frequency f . It is to

be noted that there is no delay at the first microphone as it is taken as the reference.

Page 39: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 18

Considering the frequency dependent filter coefficients w (f), the beam-former response

with respect to source direction θ is given as [63]:

r (θ, f) = w (f)H d (θ, f) . (2.18)

The weights of the beam-former are determined such that the desired source adds

constructively at the output [64]. Source signals from other than the desired source

location are filtered due to different time lag at each microphone.

It is a well known fact that the sampling rate must be at least twice the highest

frequency component of the signal to avoid aliasing, as dictated in the Nyquist theorem

[65]. It is interpreted as a scenario in which at least two points of the maximum

frequency component are being sampled. This fact is also transferable to microphone

arrays [66]. To avoid spatial aliasing a minimum distance, d, is to be maintained

between the microphones such that following inequality holds [49][63]:

d <λmin

2. (2.19)

Here λmin is the smallest wavelength that can be sampled correctly while utilising an

array with d distance between microphones.

Beam-forming algorithms can be sensitive to errors in the array characteristics [67].

Undesired correlation from one sensor to another pass through the beam-former like

white noise. Hence, gain against white noise is a measure of robustness [67]. In addition

to improving the spatial response, weight coefficients are adjusted to vary the White

Noise Gain (WNG) of beam-former. Given by:

WNG = 10 log10 |wHw|. (2.20)

Higher WNG corresponds to greater amplification of the random noise [67]. Therefore,

WNG needs to be taken care of along with the spatial response, as the spatial response

might be perfect but still WNG could lead to ineffective beam-forming.

The beam-forming techniques operate in frequency domain by making use of a narrow

band beam-former at each frequency bin [63]. This can lead to a high computational

Page 40: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 19

cost due to weights assigned to each frequency bin for achieving the required delay

and attenuation to perform source separation. As such, sub-band techniques of beam-

forming have been proposed to overcome this problem to an extent [68].

2.1.4 Sparsity-based Methods

The time-frequency sparsity of sound sources is exploited by some techniques to per-

form source separation. In this section time-frequency masking is introduced and the

most commonly used sparsity-based technique, Degenerate Unmixing Estimation Tech-

nique (DUET) is discussed [27]. Musical noise associated with such techniques is also

discussed in this section.

2.1.4.1 Time-Frequency Masking

To implement time-frequency masking the basic underlying assumption is that the

source signals are W-disjoint orthogonal. The condition of W-disjoint orthogonality is

mathematically represented as [65]:

si (f, t) sj (f, t) = 0 ∀i 6= j, (2.21)

where si (f, t) represents data of i-th source and sj (f, t) of the j-th source. It implies

that at any time frame, t, only one source is present at a certain frequency f . This

assumption holds true for a small number of sources in anechoic conditions. As the

number of sources increases, with increasing reverberations the assumption loses valid-

ity as the time-frequency data between the sources starts to overlap, although source

separation can still be performed utilising approximate W-disjoint orthogonality [69].

To achieve source separation, a time-frequency mask is generated and it is multiplied

with the microphone mixture signal [65]:

yi (f, t) = mi (f, t)x (f, t) . (2.22)

Here x (f, t) represents the microphone data, mi (f, t) represents the mask used to

separate the i-th source and yi (f, t) represents the separated output. The goal is to

Page 41: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 20

design a time-frequency mask such that the source of interest is separated without

distortion while suppressing the interfering sources. Some masks are binary (consisting

of 0s and 1s), while others are soft as they vary slowly between the minimum and

maximum value.

2.1.4.2 Degenerate Unmixing Estimation Technique (DUET)

DUET makes use of two microphones to calculate the relative amplitude and delay for

each time-frequency component and works in the underdetermined case when domi-

nant reflections and reverberations are less [27]. It requires less computational cost

as compared to the statistical and adaptive beamforming methods [27]. Due to such

operational speed, it is capable of dealing swiftly with moving sources [24]. The relative

amplitude and delay are mathematically given by [70]:

(ai (f, t) , δi (f, t)) =

[∣∣∣∣X2 (f, t)

X1 (f, t)

∣∣∣∣ ,(−1

ω

)arg

(X2 (f, t)

X1 (f, t)

)]. (2.23)

Here ai (f, t) and δi (f, t) are the relative amplitude and delay difference between the

microphone signals. Clustering is done for amplitude and delay values of each frequency

bin. With a known number of sources, standard techniques such as K-means can be

used [71], on the other hand for unknown number of sources, the number of clusters

need to be estimated first. Once the clustering is achieved, binary masks are applied

as:

m (f, t) =

1 if (a (f, t) , δ (f, t)) ∈ Gi

0 if (a (f, t) , δ (f, t)) /∈ Gi,

(2.24)

where Gi represents the i-th cluster. If a certain time-frequency component falls into

the cluster of source of interest, it is allowed to be part of the output, otherwise it is

suppressed. Such time-frequency components are converted back into the time domain

by inverse Fourier transform to achieve the final separated source output.

DUET has been used to imitate the human hearing system by mounting two micro-

phones on the sides of a dummy head [10]. The use of dummy head gives more fo-

cused clusters than conventional DUET. More focused clusters are a result of increased

Page 42: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 21

discrimination between time-frequency components due to the dummy head transfer

function. Another approach utilises multiple pairs of microphones and performs clus-

tering in the multidimensional space [15]. As such, more microphones are used and

the computational cost increases; however, it results in greater robustness and higher

resolution.

2.1.4.3 Noise Associated with Sparsity-based Time-Frequency Techniques

Among previously discussed techniques, ICA attempts to inverse the mixing matrix

and beam-forming relies on microphone array geometry to perform filtering in spatial

domain. Sparsity-based techniques, although computationally inexpensive, introduce

musical noise by attenuating the time-frequency components of mixture signals. Musi-

cal noise is heard when an output has been separated utilising time-frequency masking

of the isolated source peaks [72]. Denoising such musical noise can be a difficult task

[73]. It is commonly observed that a trade-off exists between attenuating the time-

frequency components of interfering sources and the musical noise. As a result, atten-

uating interfering sources leads to more musical noise in the source of interest and vice

versa.

2.1.5 Moving Sources

The previous sections discuss stationary source scenarios where the mixing model is as-

sumed to be time-invariant. In practical scenarios however, when sources are physically

moving, the mixing model becomes time-variant [74]. Added to it the varying room

reflections and reverberations, the scenario is further complicated to perform source

separation. To compensate for these changing mixture model parameters, a source

separation system needs to adapt. Mixing model parameters can be extracted from the

time-variant RIR estimation in order to detect, track and deconvolve moving sources

over short durations of time.

ICA-based algorithms require several iterations for deriving the unmixing matrix, lead-

ing to higher computational costs. As a result, much lighter implementations were

Page 43: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.1. Related Work 22

proposed to reduce the computational costs, termed as FastICA [75]. Even after suc-

cessful separation of the source signals using ICA, the permutation ambiguity arises

while reconstructing source signals from the separated components at each frequency

bin. The permutation problem adds to the computational cost as post processing steps

need to be carried out [55][51]. For ICA, commonly two assumptions are made to

consider moving source separation. First the sources are assumed to be physically sta-

tionary over very short time durations [76][62][77][78]. The second assumption is based

on the piecewise linearity of mixing model parameters [79]. The statistics of the signals

are not accurately measured over short time durations, leading to calculation of an

incorrect global minimum in the unmixing process. On the other hand, much longer

periods of time cause problems for ICA as Gaussianity increases with increased data

[55]. There exists a solution to improve the performance over short time durations, but

only when the number of weights to be trained is low [79].

In more recent work on moving speakers using ICA, visual cues have been used to

help solve the permutation problem by exploiting data from 3D trackers [60]. The

beam-forming and intelligently initialized FastICA are used together as a result of

visual source tracking to perform source tracking and separation [60]. In other related

work with ICA, semi-blind source separation is performed using prior knowledge of

the cancellation filters (CF) for a target region in space [23]. Fundamentally, due to

data length limitations, the stochastic algorithms alone cannot adequately estimate the

time-varying unmixing filter coefficients for moving speaker convolutive mixtures in a

blind manner [60]. Only a handful of publications can be found in the field of BMSS,

either assuming instantaneous mixtures or assuming sufficiently slowly varying mixing

coefficients [80][81][82][62][77][79], and almost all of this work has been done considering

two speakers.

A popular technique among sparsity-based algorithms is DUET [27]. It exploits relative

amplitude and delay difference between microphones to cluster similar time-frequency

components. More recent work utilising DUET employs Expectation Maximization

(EM) initialised by RANdom SAmple Consensus (RANSAC) to cluster the wrapped

interchannel phase differences [24]. For moving speakers with RANSAC preprocessed

data, the factorial wrapped Kalman filter has been used to perform speaker tracking

Page 44: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 23

[24]; however, the number of speakers are known prior to performing separation with

the assumption that the target source is always active, which is practically not very

viable.

Work related to source detection and tracking has been done using the combination of

DUET and cardinality-balanced multi-target multi-Bernoulli filter, however simulated

data with two source mixtures has been considered without performing source sepa-

ration [83][17]. Also the steered power response (SPR) is cited frequently for sound

source localisation applications [84][85][86]; however, a direct implementation of SRP

for the time-variant mixing conditions requires excessive computations, increasing the

waiting time of processing to even twenty-four hours for one minute of the recording

on a desktop PC [87].

Theoretically, moving source separation can be performed by using inverted RIRs, given

the source locations are constantly updated, as the room mixing coefficients are hidden

in the room impulse responses [5]. With such knowledge the corresponding unmixing

parameters can be calculated to perform source separation. If the source locations are

not accurately updated in a reverberant environment, small changes in source location

may change the mixing system parameters drastically, leading to poor performance.

Therefore with such room exploitation, source separation is currently infeasible.

Array oriented techniques are more appropriate for dealing with moving sources, pro-

vided the system is capable of swift adaptation and the updated source locations are

accurate. The sparsity-based techniques can provide with an elegant low complexity

solution to the moving source separation problem. The problem lies with obtaining

accurate source locations and swiftly reacting to the changing mixing parameters. Fur-

thermore to achieve a blind source separation system, source detection is also critical to

detect time varying number of sources as they enter or leave the dynamically changing

acoustic scenario.

2.2 Intensity Vector Direction based Source Separation

The time-frequency sparsity-based technique proposed by Gunel et al. [11] utilises

Page 45: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 24

the intensity vector direction statistics to perform source separation. The IVD-based

source separation algorithm provides a robust and nearly closed-form solution to the

source separation problem. The mixtures are captured by a compact, nearly coincident

microphone array, which is very desirable in small electronic devices such as cell phones

or hearing aids, and the algorithm is capable of performing source separation in real-

time. In this section the background of IVD-based system with respect to microphone

array and IVD measurement is detailed.

2.2.1 Spatial Audio and B-format

The most basic form of audio capture is known as mono, which contains information

in a single track. Mono is capable of conveying only some spatial information, such as

reverberation or depth. For example the echo effect of a big hall can be perceived in

a single channel, but it lacks in giving a sense of direction. To convey the directional

information, at least two audio tracks or added frequency effects in the mono audio

are required [88]. The form of audio which uses two audio tracks to convey a sense

of direction is known as two channel stereo. Stereo sound uses two or one of the

two means to reproduce sound with a sense of direction. The first one is the timing

difference between two tracks, and the second one is level difference. These are termed

as the inter-aural time difference (ITD) and the inter-aural level difference (ILD). These

differences create a sense of direction when the sound is reproduced. Only one of these

two means may also create the sense of direction, but the change has to be a major

one in that case. For example, if only inter-aural time difference is used, the time

difference has to be larger than it would require for both the cues to give the same

sense of direction [89].

The sense of direction due to ITD and ILD difference is used in the spatial sound

reproduction as well. More advanced form of spatial sound reproduction, which is very

common these days, is known as 5.1 channel [90]. The term 5.1 refers to five tracks

of audio played around the listener, and a single low frequency track. Over the years

with more direction cues, the number of audio tracks has increased to 22.2 [91]. The

principle, however, of sound reproduction is the same as that of stereo. The ITD and

Page 46: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 25

ILD in the sound belonging to different tracks played around the listener create a sense

of spatial submergence.

The ambisonics is also a form of spatial audio which does not feed the speaker system

directly, but it carries the directional information of an entire sound field [88]. A decoder

is designed to produce spatial sound for the desired speaker layout. The method for

capturing 3D spatial information using the coincident configuration of the microphones

was proposed in [92]. The coincident configuration implies microphones at the same

location having different directivity patterns. The quality of directional information

encoded in the coincident microphone recordings is dependent on arrangement of the

microphones and their directivity patterns. The microphone responses are categorized

in orders. Zero-th order microphones, such as the omnidirectional microphones, are

equally sensitive to all the directions of arrival [93]. The response of the first order

microphones, however, depends on the direction of arrival of sound. The figure-of-

eight microphones are first order microphones, picking up sound from the front and

the rear, but not from the sides [93]. The microphone responses up to first order

are discussed in this section, these are distinct from the higher order responses [94].

The first order microphones with different directional characteristics can be realized by

additive combination of the coincident omnidirectional and bidirectional microphones:

f (θ) = ζ + (1− ζ) cos (θ) (2.25)

here ζ, varying from 0 to 1, determines the contribution of the omnidirectional and

the bidirectional component to the combination. The cosine function represents the

bidirectional response and θ is the direction of arrival of sound. Some ideal first order

directivity functions are shown in Fig. 2.3.

2.2.2 B-Format Audio for IVD Measurement

In this section the directivity patterns are assumed to be theoretical i.e. they have the

same response for all frequencies. In the practical domain, this assumption is no longer

valid as the directivity patterns are deformed at very low and high frequencies [95][96].

Page 47: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 26

Figure 2.3: First order microphones showing the sensitivity towards directions of arrival

in the surrounding space (range of sensitivity = 0-1). Figure-of-eight (solid line ζ = 0),

Sub cardioid (dashed-dotted line ζ = 0.7), Cardioid (dashed line ζ = 0.5)

The first order B-Format audio, captured using a coincident microphone array [92],

comprises of the three bidirectional responses and one zero-th order omnidirectional

response. These signals collectively describe the sound field as a three dimensional

entity, at a certain point in space [97]. The B-format audio is also used in commercial

recording applications, as it provides some useful functionalities over conventional sound

recording techniques. One major functionality is that the recorded B-format audio can

be converted into any desired 3D multichannel audio format [98], such as two channel

stereo or 5.1.

Generally many sources are active in the surrounding space, hence the B-format signals

represent the superposition of their individually generated pressure and pressure gra-

dient signals. As such, the directional information carried by the first order B-format

audio is exploited to perform source separation by carrying out IVD measurements [11].

For a two dimensional scenario the B-format audio is composed of WB, XB, and YB

audio channels. These are mathematically represented as [99]:

WB(ω, t) = 1/√

2× P (ω, t), (2.26)

XB(ω, t) = cos(θ)× P (ω, t), (2.27)

Page 48: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 27

YB(ω, t) = sin(θ)× P (ω, t), (2.28)

where θ is the horizontal angle of the incoming acoustic wave at a particular time-

frequency with respect to the microphone array. WB(ω, t) represents the time-frequency

omnidirectional pressure. XB(ω, t) and YB(ω, t) represent time-frequency pressure gra-

dients along the x axis and the y axis respectively. The pressure gradient measurements

help to measure the projection of the source pressure along the x and y axis, leading

to a vector analogy. The first order B-Format audio signals are depicted in Fig. 2.4.

Figure 2.4: First order B-format Audio [100]

Using B-format audio channels, the intensity vector measurements along the horizontal

plane are carried out as [11]:

~I (ω, t) =1

2ρc(Re {W ∗

B (ω, t)XB (ω, t)}+ i (Re {W ∗B (ω, t)YB (ω, t)})) . (2.29)

The Intensity Vector Direction (IVD) can hence be measured using equation (2.29) as

[11]:

IV D = tan−1

Im{~I (ω, t)

}Re{~I (ω, t)

} , (2.30)

where Im and Re represent the imaginary and real components respectively.

Similarly the energy of each time-frequency component arriving from a particular lo-

cation in space is measured using B-format signals as [101]:

E (ω, t) = 0.5(WB (ω, t)2

)+ 0.5

(XB (ω, t)2 + YB (ω, t)2

). (2.31)

Page 49: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.2. Intensity Vector Direction based Source Separation 28

Figure 2.5: Tetrahedral placement of microphones to form A-format microphone array

2.2.3 Microphone System for IVD Measurement

The required pressure and pressure gradient signals are captured with the help of

a tetrahedral microphone array, as shown in Fig. 2.5. It is generally achieved by

mounting four cardioid microphones on the faces of a regular tetrahedron [92]. The

regular tetrahedral positions also imply non-adjacent ends of a cube. The four sig-

nals will be attributed as Left-Forward-Up, Front-Right-Down, Back-Left-Down and

Back-Right-Up, with reference to centre of the tetrahedron. These are collectively

termed as A-format signals. The A-format signals are contained in a vector Af (t) =

[PLFU (t) PFRD (t) PBLD (t) PBRU (t)], where PLFU (t) represents the pressure signal

captured by microphone at position Left-Forward-Up at time t. Likewise PFRD (t),

PBLD (t) and PBRU (t) represent the Front-Right-Down, Back-Left-Down and Back-

Right-Up pressure signals respectively. For conversion to B-format signals, the process

carried out can be mathematically represented in the matrix form as [97]:

B(t) = FAf (t). (2.32)

F =

+1 +1 +1 +1

+1 +1 −1 −1

+1 −1 +1 −1

+1 −1 −1 +1

. (2.33)

Page 50: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

2.3. Conclusion 29

The signals contained in B (t) represent the omnidirectional pressure and the pressure

gradients along the x, y and z axes. These are mathematically represented for the

horizontal scenario in equation (2.26), equation (2.27) and equation (2.28). Once the

B-format signals have been formulated, IVD measurements can be carried out using

equations (2.29) and (2.30).

2.3 Conclusion

Source separation algorithms can be widely grouped into three distinct categories.

These include stochastic, beam-forming and time-frequency sparsity based source sep-

aration algorithms. Stochastic algorithms such as ICA iteratively find unmixing filter

parameters assuming statistical independence of source signals. Beam-forming methods

rely on array geometry to separate the desired source signal in space while exploiting

different times of arrival of incoming signal at each microphone. On the other hand

deterministic techniques, such as DUET or IVD based system, utilise time-frequency

sparsity among source signals to perform separation. This thesis deals with the analysis

and improvement of IVD based source separation technique which uses level difference

among closely spaced microphones. It has several advantages over other source sep-

aration techniques such as ICA or DUET. The IVD-based algorithm uses a compact

and nearly coincident microphone array to exploit multipath characteristics of the sur-

rounding environment. The number of sources to be separated is not restricted by the

number of microphones, and the algorithm does not need to converge to unmixing filter

parameters to perform separation. The IVD measurements can be carried out using an

acoustic vector sensor or a commercially available A-format microphone array.

Page 51: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Chapter 3

Improvement in IVD-based

Source Separation for Stationary

Sources

Preview

This chapter presents analysis and algorithms for improvements to an intensity vec-

tor direction (IVD) based audio source separation algorithm. The aim is to improve

the applicability and output sound quality of IVD-based system. Based on the given

source locations, we propose an adaptive system that blindly caters for the effect of

environment and other interfering sources through energy weighted spatial filter pa-

rameters. It was investigated that due to physical constraints of the microphone ar-

ray, distorted space-frequency response affects the broadband source localisation cues.

A measurement-based algorithm is proposed to improve on such microphone array

shortcoming. Furthermore, microphone array weighting is proposed to select the most

relevant cardioid microphones for desired source locations. As a result of proposed

modifications, the system is capable of adapting to suitable filter parameters in a non-

exhaustive manner and provides significantly better output sound quality. Improvement

in numerical evaluation for two rooms with two and four source combinations of speech

and music is reported. As compared to state-of-the-art methods, the IVD-based system

30

Page 52: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.1. Introduction 31

has an advantage of employing a small microphone array. It provides a deterministic

and nearly closed-form solution to the source separation problem, with the capability

of performing source separation in real-time. First-hand impact of the IVD-based sys-

tem includes 3D audio signal processing, noise suppression, teleconferencing, acoustic

surveillance, high quality hearing aid and speech and speaker recognition applications.

3.1 Introduction

As presented in detail in Section 2.1, the source separation techniques can be grouped

into three distinct categories. These are the stochastic, deterministic and adaptive tech-

niques. The stochastic methods, such as those based on independent component anal-

ysis (ICA), perform separation by assuming statistical independence amongst source

signals [7][8][35][102]. ICA was initially proposed for the separation of sources from

instantaneous mixtures. As convolution in the time domain corresponds to multipli-

cation in the frequency domain [103], ICA was modified to carry out implementation

in parallel on each frequency bin to deal with convolutive mixtures. ICA-based algo-

rithms are computationally expensive as they require several iterations for deriving the

unmixing filter coefficients. As a result much faster implementations were proposed,

such as FastICA [45]. The FastICA algorithm is robust, computationally light and

converges very fast as compared to conventional ICA algorithms [104]; however, it still

needs to converge to a solution [105][75]. In addition, frequency domain ICA-based

techniques suffer from scaling and permutation ambiguities [55]. After finding the in-

dividual source components at each frequency bin, these algorithms need to assign

them to their respective sources [106][51][107][108][109]. The post processing required

to solve the permutation problem adds to the computational cost. There is no closed

form solution to solve the permutation ambiguity, as such it is still being investigated

[58][53].

Adaptive algorithms, such as Adaptive Beam-Forming (ABF), perform optimization ac-

cording to the source signal properties in order to derive multichannel unmixing filters

[11]. ABF is dependent on microphone array geometry and has the ability of spa-

tial selectivity to perform separation of the desired source while suppressing unwanted

Page 53: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.1. Introduction 32

interfering sources [3][4]. The adaptive and stochastic algorithms are dependent on

source signal properties to converge to unmixing parameters; however, the stochas-

tic algorithms do not need the knowledge of the microphone array geometry to their

advantage.

The sparsity-based deterministic algorithms, such as the Degenerate Unmixing Estima-

tion Technique (DUET), are fundamentally different from the stochastic and adaptive

algorithms. Instead of the source signal properties, DUET exploits multipath character-

istics of the reverberant environment by utilising the relative delay and amplitude differ-

ence between two microphone signals [65][27]. The DUET has been extended to more

than two microphones using the k-means clustering algorithm, termed as MENUET

(Multiple sENsor dUET) [15]. More recent work utilising DUET employs Expectation

Maximization (EM) initialised by RANdom SAmple Consensus (RANSAC) to cluster

the wrapped inter-channel phase differences [110].

A Model based Expectation maximisation Source Separation and Localisation (MESSL)

algorithm has been proposed by Mandel et al. [111] which uses two mixture channels.

Time-frequency components clustered using inter-aural phase and inter-aural level dif-

ferences are used to generate isolation masks for individual sources. A probabilistic

mixture model of the inter-aural parameters is exploited at each time-frequency unit

to facilitate source separation. Alinaghi et al. [112] proposed fusion of the inter-aural

phased and inter-aural level difference with mixing vector distributions to achieve a hy-

brid source separation algorithm which also uses two mixture channels. Mixing vector

distributions provide independent representations when the sources are close to each

other. On the other hand, inter-aural parameters are more robust to high reverber-

ation time than the mixing vector models. An added advantage of using inter-aural

parameters is that the permutation problem of frequency domain BSS is addressed by

initialisation of mixing vectors based on inter-aural parameters [112].

The IVD-based technique proposed by Gunel et al. [11] also utilises the space-frequency

sparsity to perform source separation. Considering the processing time of source sepa-

ration algorithms, it is very desirable to have a deterministic closed-form solution. The

IVD-based algorithm proposed by Gunel et al. [11] provides a nearly closed-form solu-

Page 54: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 33

tion to the BSS problem of convolutive mixtures. The algorithm performs separation

of the desired source by utilising given information in the form of source locations, and

exhaustive calculation of the spatial filter beam-widths [11]. The mixtures are recorded

by a compact, nearly coincident microphone array, which is very desirable in small elec-

tronic devices such as high quality hearing aids and cell phones. The number of sources

to be separated is not restricted to the number of obtained mixtures [11]. Prior to

Gunel et al. [11], no similar method using a nearly coincident microphone array with

satisfactory performance had been proposed [11]. In almost all other source separation

techniques, such as based on ICA, adaptive beamforming or DUET, non-coincident mi-

crophone arrays are used [11], making the IVD-based source separation system distinct

in its approach. From the earlier literature, it is found that coincident microphone

arrays have been mostly investigated for sound source localization, directional audio

coding and intensity vector calculations [113][114][101].

The motivation behind work presented in this chapter is to improve the baseline IVD-

based source separation system in terms of output sound quality and applicability.

The baseline IVD-based source separation system is discussed in Section 3.2. Section

3.3 presents analysis of the baseline IVD-based system to examine and understand its

behaviour and potential with respect to different aspects. Section 3.4, 3.5, 3.6 and 3.7

detail the proposed system methodology along with its evaluation. Section 3.8 describes

the experimental test conditions and presents the obtained results in comparison to the

baseline IVD-based system, and Section 3.9 concludes the chapter.

3.2 Baseline IVD-Based Source Separation System

In an environment where various sound sources are active, the omnidirectional pressure

signal WB(ω, t) of n-th source is given by [11]:

WB(ω, t) =

N∑n=1

sn(ω, t)hn(ω, t), (3.1)

where sn(ω, t) represents n-th source signal in the time-frequency domain. The time-

frequency representation of transfer function from the n-th source to microphone is

Page 55: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 34

DFT

IVDMeasurement

A-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

ŝn (t)

Spatial Filters

IDFT

Source Locations

W(ω,t)

Figure 3.1: Processing stages of the baseline IVD-based system [11]. Initially, the pres-

sure and pressure gradient signals are obtained from the microphone array for intensity

measurements. Then, the DFT of the signals is measured. Next, the intensity vector

directions (IVD) are calculated in time-frequency domain. Next, the von-Mises spatial

filters are applied for the known source locations. Finally, IDFTs of the separated

signals are calculated to get the time-domain estimate of separated sources.

hn(ω, t). Based on the desired source location, source separation is performed by es-

timating the contribution of source sn(ω, t) to the omnidirectional mixture WB(ω, t)

provided that hn(ω, t) is unknown.

Gunel et al. [11] proposed to tackle the source separation problem by estimating the

IVD using intensity measurement of each time-frequency unit. The intensity measure-

ments can be carried out by using closely spaced microphones to measure the pressure

and pressure gradient signals [115]. After IVD measurements, soft masking using spa-

tial filters is carried out to filter the desired source components. In a reverberant

environment the spatial filters are modelled as von-Mises distribution functions [11].

The von-Mises distribution function is circular counterpart for the well known Gaussian

distribution function. The von-Mises distribution function is given by [116]:

f (θ;µ, σ) =eσ cos(θ−µ)

2πIo (σ), (3.2)

Page 56: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 35

where µ is the mean circular direction, θ is a circular variable lying between 0 and

2π. The concentration parameter σ is similar to the variance of Gaussian distribution.

Io (σ) is the modified Bessel function of zero order. Here σ is logarithmically related to

the 6 dB beam-width θBW as [11]:

σ =ln 2

1− cos(θBW2

) . (3.3)

The mean circular direction µ is given in the form of desired source location, while the

suitable concentration parameter σ is found by exhaustively searching for the optimum

θBW [11]. For the IVD histogram, best fitting θBW is found by varying the θBW

from 0 degrees to 180 degrees in 10 degree intervals [11]. The fitting gives ineffective

results at less angular intervals and the exhaustive search for optimum beam-width is

computationally expensive [117]. The von-Mises spatial filters with varying circular

mean and concentration parameters are depicted in Fig. 3.2.

Figure 3.2: The von-Mises spatial filters used for soft masking the time-frequency

components based on their IVD measurements. The spatial filters are depicted with

varying circular means and concentration parameters as mentioned in the legend of the

figure.

Page 57: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 36

z

x

y

LFU

RFD

LBD

RBU

r

Figure 3.3: The A-format microphone array with tetrahedral configuration of the car-

dioid microphones: Labelled as Left-Forward-Up (LFU), Right-Front-Down (RFD),

Left-Back-Down (LBD) and Right-Back-Up (RBU) with respect to the centre of the

configuration. Collectively these are known as the A-format signals. The distance of

the microphones to center of the array is denoted by r.

3.2.1 Pressure and Pressure Gradient Measurements

The intensity measurements can be carried out through the pressure and pressure gra-

dient signals [115]. To acquire pressure and pressure gradient signals the microphones

are positioned close together to avoid aliasing at high frequencies [95][115]. The co-

incident microphone arrays, such as the A-format, can be utilised to have the desired

pressure and pressure gradient measurements because they are assembled to capture

the sound field at the same point in space [118][95][11][96]. The A-format array consists

of four closely spaced cardioid, subcardioid or omnidirectional microphones, placed at

the non-adjacent ends of a cube to capture the 3-dimensional sound field [119] (see Fig.

3.3). The A-format microphone array and the Acoustic Vector Sensor (AVS) [120][121]

have been used in previous works for sound intensity measurements in order to perform

source separation. Although any pressure and pressure gradient measuring microphone

array can be used, mostly the commercially available arrays are utilised. As such, the

microphone arrays manufactured by Core Sound [122] or Soundfield [119] are among

Page 58: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 37

the available options.

3.2.2 Frequency Domain Representation

While analysing the sound field for source separation using the IVD measurements,

a sparse time-frequency representation of the sources is obtained through short time

Fourier transform (STFT) [11]. The underlying assumption while carrying out the

transformation into time-frequency domain is that the sources are W-disjoint orthog-

onal. The W-disjoint orthogonality condition is fulfilled when only a single source is

present at a particular time-frequency unit, such that [65]:

si(ω, t)sj(ω, t) = 0, (3.4)

where si(ω, t) is the time-frequency content from i-th source and sj(ω, t) is time-

frequency content from j-th source. To carry out STFT processing, the A-format

signals are windowed via a sin function [11]:

wk = sin

2sin2

2M

(k +

1

2

)]), (3.5)

where 2M is the window size in samples and k is the sample index. If the source

signals are not W-disjoint orthogonal, the same time-frequency unit will be occupied

by multiple sources simultaneously, leading to inaccurate IVD measurement of that

time-frequency unit [11]. In reality the source signals are approximately W-disjoint

orthogonal [110], and as the number of sources increase the W-disjoint orthogonality

is violated to a greater extent [65]; however, the sources are still fairly W-disjoint

orthogonal to achieve adequately sparse time-frequency representations [11][24].

3.2.3 A-format to B-format Conversion

The pressure and pressure gradient signals can be directly captured using a commer-

cially available B-format microphone system, such as the SoundField SPS422B micro-

phone system [119]; however, with the aforementioned A-format microphone array, the

Page 59: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.2. Baseline IVD-Based Source Separation System 38

signals are converted into omnidirectional pressure and bidirectional pressure gradi-

ent signals through summations and subtractions [119]. The resultant omnidirectional

pressure and bidirectional pressure gradient signals are collectively termed as B-format

signals [97]. The B-format signals are achieved as [97]:

WB(ω, t) = PLFU (ω, t) + PRFD(ω, t) + PLBD(ω, t) + PRBU (ω, t) (3.6)

XB(ω, t) = PLFU (ω, t) + PRFD(ω, t)− PLBD(ω, t)− PRBU (ω, t) (3.7)

YB(ω, t) = PLFU (ω, t)− PRFD(ω, t) + PLBD(ω, t)− PRBU (ω, t). (3.8)

Here PLFU , PRFD, PLBD and PRBU are the left-forward-up, right-forward-down, left-

back-down and right-back-up A-format pressure signals respectively, with the centre of

the cube as reference (see Fig. 3.3). The WB(ω, t) signal represents omnidirectional

pressure and, XB(ω, t) and YB(ω, t) represent the pressure gradient signals along x-axis

and y-axis respectively. A pictorial representation of WB, XB, and YB B-format signals

is presented in Fig. 2.4.

3.2.4 IVD measurement

With pressure and pressure gradient signals, in the form of B-format, for each time-

frequency unit the IVD measurement along the horizontal plane can carried out as [11]:

~I (ω, t) =1

2ρc(Re {W ∗

B (ω, t)XB (ω, t)}+ i (Re {W ∗B (ω, t)YB (ω, t)})) (3.9)

IV D = tan−1

Im{~I (ω, t)

}Re{~I (ω, t)

} , (3.10)

where ~I (ω, t) represents the intensity vector measurement and Im and Re represent

imaginary and real components respectively.

The IVD measurement is dependent on several factors including microphone array

response, reverberant response of the surrounding environment, the impact of multiple

speakers on a particular time-frequency unit and observed finite aperture of the sound

source [11]. As a result, while analysing the intensity vector directions in space, a sharp

clustering around the source locations is not observed [11].

Page 60: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 39

3.2.5 Spatial Filtering

Ultimately, in the baseline IVD-based system, source separation is achieved by soft

masking in time-frequency domain to filter out IV D (ω, t) with the help of source spe-

cific von-Mises spatial filters. The time-frequency estimate of the n-th source, sn (ω, t),

can be obtained as:

sn (ω, t) = WB (ω, t) f(θ;µn(t), σn(t)). (3.11)

Here WB (ω, t) is the omnidirectional pressure signal, f(θ;µn(t), σn(t)) represents the

von-Mises spatial filter for the n-th source, and for N sound sources N von-Mises spatial

filters are utilised [11].

3.2.6 Time Domain Representation

At the final stage, after different source components have been identified and filtered

within each STFT frame, the inverse discrete Fourier transform (IDFT) is calculated

to have the time domain estimates of separated sources as sn(t) [11].

3.3 Analysis of the IVD-based System

In this contribution the baseline IVD-based system will be investigated with different

perspectives. Subsequently, solutions with low computational costs will be proposed in

order to improve in areas where drawbacks are identified.

Owing to the fact that practical digital signal processing systems may deviate from the-

oretical assumptions due to imperfections and undesirable components in the acquired

data [123], in this section the baseline IVD-based system is analysed and solutions are

discussed for the observed shortcomings. First, the objective performance of the base-

line IVD-based system is presented with varying angular intervals between the sources

and varying spatial filter parameters. After that the source separation system is anal-

ysed from three different perspectives, firstly with respect to the microphone array,

secondly with respect to the surrounding environment and thirdly with respect to the

impact of multiple sources on the desired source.

Page 61: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 40

3.3.1 Objective Performance Evaluation

Before analysing the system with different perspectives, the baseline IVD-based system

is analysed with objective performance measures to help understand the impact of

spatial filter parameters on the source separation performance. The experimental set-

up and the objective measures are discussed before the results and discussion section.

3.3.1.1 Obtaining the Mixtures

The convolutive mixtures used for testing the baseline IVD-based system and later the

proposed algorithms were obtained by measuring A-format room impulse responses.

The impulse responses were convolved with anechoic sound sources and summed in de-

sired combinations to get reverberant source mixtures. This approach utilises linearity

and time-invariance assumptions of linear acoustics as carried out by Gunel et al. [11].

The impulse responses were measured in two different rooms using the sine-sweep tech-

nique [124]. The first room was an ITU-R BS1116 standard listening room with a

reverberation time (T60) of 0.32s. The second room was being utilised as a vision lab

with a reverberation time of 0.67s. For both the rooms, 36 A-format impulse response

recordings were obtained at 44100 Hz with Core Sound TetraMic [122] and a loud-

speaker (Genelec 8010A). Each of the 36 measurement positions were located on a

circle of 1.5m radius for both the rooms, and the microphone array was placed at the

center of the circular set-up. The source locations were selected with respect to the

microphone array in between 0 and 350 degrees, with 10 degree intervals. At each

measurement position, the acoustical axis of the loudspeaker was facing towards the

microphone array, while the orientation of the array was fixed. The sources and the

microphone array were 1.3m high above the floor to demonstrate source separation in

the horizontal plane.

Anechoic sources sampled at 44100 Hz were used from the Music for Archimedes com-

pact disk, recorded as part of the European project named Archimedes [125]. The

5s long portions of male English speech (M), female English speech (F), cello music

(C), and trumpet music (T) were used. They were convolved with A-format impulse

Page 62: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 41

responses in the desired source and location combinations. The A-format sounds were

then summed to obtain CT, MT, MC, FT, CF, and FM for two source mixtures and

TCFM, CFMT, FMTC and MTCF for four source mixtures with different angular

configurations.

In this section results for the ITU-R BS1116 standard listening room (T60 = 0.32s)

with two source mixtures have been presented for discussion.

3.3.1.2 Objective Measures

A widely used blind source separation performance toolbox, BSS Eval [126], was used

to assess the output separated signals. The toolbox performs evaluation objectively by

estimating the break down of separated source signal into the target source components,

interference components, noise components and other artefacts. With the help of this

breakdown SDR is calculated as: [126]

SDR = 10 log 10Starget

2

(einter + enoise + eartef )2. (3.12)

The SDR measure depends on overall physical characteristics of the signal; however,

alone it does not justify the separation quality completely, especially when angular

intervals are less. The Signal to Interference ratio (SIR) was utilised for analysing the

interfering source components independently. The SIR measure is given as [126]:

SIR = 10 log 10Starget

2

einter2. (3.13)

The reference signal was chosen as a static cardioid derived from B-format signals,

directed towards the target source location [11].

3.3.1.3 Results and Discussion

The average SDR and SIR results for two source separation with varying angular in-

terval and varying von-Mises spatial filter beam-width is presented in Fig. 3.4 and Fig.

3.5 respectively. At a greater angular interval between the sources, the separated

Page 63: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 42

Figure 3.4: The SDR (dB) results of varying angular interval between the sources at

different beam-widths of the von-Mises spatial filters. The black line is indicating the

trend of increasing SDR as the optimum line.

Figure 3.5: The SIR (dB) results of varying angular interval between the sources at

different beam-widths of the von-Mises spatial filters. The black line is indicating the

trend of increasing SIR as the optimum line.

Page 64: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 43

source quality improves as the spatial filter beam-width increases; however to a certain

extent. Beyond a limit, the greater beam-width captures significant amount of the

interfering source components and the separation performance degrades.

In Fig. 3.4 and Fig. 3.5 at low angular intervals, lesser beam-widths can be seen to

be more effective as source energy distributions start to overlap in space as the sources

come closer. In such a scenario, to separate the sources effectively, a low spatial filter

beam-width is required. This is represented by the optimum line shown on the SDR

and SIR plots, as it passes through low beam-width values for low angular intervals (20

degrees to 60 degrees).

20 30 40 50 60 706

7

8

9

10

11

Beam−width

SD

R (

dB

)

(a) SDR: RT = 0.32s

20 30 40 50 60 7018

19

20

21

22

23

Beam−width

SIR

(d

B)

(b) SIR: RT = 0.32s

Figure 3.6: Average objective evaluation of two source separation by varying spatial

filter beam-width for baseline IVD system. The overall SDR performance is seen to in-

crease with the increase in spatial filter beam-width (Fig. 3.6a). With a greater spatial

filter beam-width, more source components spread around the desired source location

are captured. However, as depicted in Fig. 3.6b, consequently more components of the

interfering sources are captured as well. Due to this a trade-off exists between signal

to interference ratio and signal to distortion ratio. A suitable beam-width applicable

to all angular intervals is chosen such that SIR does not fall below a desired threshold

(approx. 20 dB).

The average performance of spatial filter beam-widths between 20 degrees and 70 de-

grees for separation throughout the varying angular interval between sources is depicted

Page 65: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 44

in Fig. 3.6 to identify the best beam-width across all angular intervals. The analysis

also shows the trade-off between SDR and SIR measures. The overall SDR performance

can be seen to increase with the increase in spatial filter beam-width (Fig. 3.6a). With

a greater spatial filter beam-width, the spatial filter is able to capture more desired

source components spread around its physical location; however, as depicted in Fig.

3.6b, consequently more components of the interfering sources are also captured as

belonging to the desired source. The aim of the source separation system should be

having optimum quality at the output while minimizing the interference from other

sources. As such, performance improvement in both the SDR and SIR measures is

sought to achieve a better IVD-based source separation system.

In light of these results, the quality of separation is dependent on angular interval

between the sources and the spatial filter beam-width. As such, the spatial filter beam-

width should have the capability of adaptation to optimum value based on the given

source locations. Detecting spatial overlap between filters directed towards different

locations can assist in adapting to low beam-widths when the angular interval between

sources is less.

3.3.2 A-format Array Characteristics

To understand the behaviour of IVD-based system it is important to analyse the mi-

crophone array being utilised for intensity measurements. In Fig. 3.3, the four A-

format microphones are shown as Left-Forward-Up (LFU), Right-Front-Down (RFD),

Left-Back-Down (LBD) and Right-Back-Up (RBU), with respect to the centre of the

microphone array. The pressure signals obtained from this configuration are collectively

termed as A-format signals. A-format signals are converted into the B-format signals

for carrying out intensity measurements. The A-format microphone array is not ideally

coincident due to its physical constraints, leading to deteriorating broadband IVD mea-

surements [96]. Even with very close positioning, there remains a small radial distance

(r), to the centre of the array (see Fig. 3.3). This non-coincidence leads to spatial alias-

ing above a certain limit frequency, fl, which is directly dependant on radial distance

r [95][96]. Greater the distance, less affective the coincidence characteristic, resulting

Page 66: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 45

in a lower limit frequency. Beyond limit frequency, the location-based space-frequency

responses of microphone array get significantly distorted [95][96]. It happens when

the wavelength of incoming sound wave becomes comparable to the distance between

microphones. The limit frequency, fl, is given in equation 3.14,

fl =c

πr, (3.14)

where c is the speed of sound (340 m/s) and r is the distance from centre of the

microphone array to effective centre of the cardioid microphones [127], as depicted in

Fig. 3.3.

Ideally, coincident microphone array should only have sound pressure level difference

among its microphone signals, but in reality they are nearly coincident. As a result, the

IVD-based algorithm is capable of having accurate IVD measurements only for a range

of frequencies [95][96]. The difference in the pressure level is determined by orientation

of the microphones and their directivity patterns. In this section the impact of A-format

microphone array size and microphone directivity patterns on the IVD measurements

is discussed.

3.3.2.1 Impact of the A-format Array Size

As the microphone array size decreases, the ability to capture higher frequencies with

accurate IVD measurements improves [128]. This is due to adequate spatial sampling

at high frequencies when the array size is small [129]. As such, for any microphone

the array size determines limit frequency, fl, as given in equation 3.14. Different array

sizes have been simulated to show this phenomenon in Fig. 3.7.

In Fig. 3.7 the IVD measurements with several different array sizes, which can be

considered as nearly coincident, are depicted. Each line originating from an angle is

representative of a location-based space frequency response of the microphone array.

As such, in Fig. 3.7, measurements of 36 locations are represented. The array size

varying from 0.6cm to 1.8cm depicts the variation in IVD measurements across the

space-frequency domain. The limit frequency can be recognized as frequency at which

Page 67: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 46

5000 10000 150000

100

200

300

Frequency (Hz)

An

gle

(D

eg

ree

s)

(a) Microphone Response r = 0.6cm (b) Microphone Response r = 1.0cm

(c) Microphone Response r = 1.4cm (d) Microphone Response r = 1.8cm

Figure 3.7: Simulation results of location-based space-frequency IVD measurements

with different microphone array sizes. The limit frequency, highlighted using red verti-

cal line, can be seen to get lower as the array size increases.

the IVD measurements from different locations start to overlap in space (see Fig. 3.7).

Beyond limit frequency, the space-frequency response of different locations is seen to

overlap due to aliasing. With a small array size of 0.6cm, as shown in Fig. 3.7a, the

limit frequency can be seen to be much higher as compared to greater array sizes.

The phenomenon shown in Fig. 3.7 indicates that the space-frequency response of

microphone array does not provide ideal localisation cues as required by the IVD-based

source separation; however, it may be improved below the limit frequency because of

spatially independent IVD measurements.

An important observation from Fig. 3.7 is the way the space-frequency IVD measure-

ments are grouped into four quadrants. The space-frequency responses are seen to

converge at 90 degrees, 180 degrees, 270 degrees and 360 degrees for different array

Page 68: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 47

(a) Top View: Source at 90 Degrees (b) Top View: Source at 45 Degrees

Figure 3.8: Different microphone behaviour depicted for sound originating from dif-

ferent locations to visualise the effect of increased non-coincidence based on source

locations.

sizes. The reason being the orientation of cardioid microphones to form the tetrahedral

arrangement. To facilitate understanding of this phenomenon, for sound originating

from two different locations, the top view of the microphone array is shown in Fig. 3.8.

The impact of sound originating from 90 degrees is depicted in Fig. 3.8a. The time of

arrival for a plane wave is same for the depicted top two microphones. The condition

of coincidence holds true for both microphones, leading to better IVD measurements

across a wider frequency range. In comparison, if the impact of sound originating from

45 degrees is visualised, as shown in Fig. 3.8b, the top two cardioid microphone sig-

nals can be seen to have different times of arrival, resulting in a poor space-frequency

response.

It is relevant to mention here that if the location-based IVD measurements of the

microphone array are known, they can be used to perform correction for the desired

source locations upto the limit frequency. As such, it is proposed later in Section

3.4 that the location-based IVD measurements of a real microphone array can act as

correction coefficients stored in the form of a look-up table.

Page 69: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 48

Figure 3.9: Different directivity patterns of the microphone facing towards 45 degree.

The sensitivity of the microphone array shown to vary from 0 to 1 for different circular

directions of arrival.

3.3.2.2 Impact of the Directivity of Microphones

The effect of omnidirectional, sub-cardioid and cardioid microphone directivity patterns

on the IVD measurements is analysed with the simulated A-format microphone array.

These directivity patterns, focusing at 45 degree, are depicted in Fig. 3.9. Despite

having different directivity patterns, with a constant array size of 1.4 cm, IVD mea-

surements remain constant, as given in Fig. 3.7c. Theoretically, the way microphones

are positioned, the resulting measurements of the pressure and pressure gradient signals

are only different in terms of scaling. As a result, the IVD measurements do not get

affected.

In practical acoustic scenarios, however, in order to have the spatial signatures of

sources embedded in the A-format signals, cardioid or sub-cardioid directivity functions

are preferred [122]. As a result, the commercial microphone manufacturing companies,

such as the Core Sound and SoundField, use cardioid or sub cardioid directivity patterns

for acquiring A-format signals before conversion to B-format [119][122]. Note that for

the cardioid weight matrix algorithm proposed later on (see Section 3.5), microphones

Page 70: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 49

Figure 3.10: The recording set-up with distances to floor and the ceiling depicted. The

first reflection was found to be approximately 3ms after the arrival time of the direct

sound component. With the speed of sound to be 340m s it comes out to be around

1m of distance from the closest reflective surface, which was the ceiling.

with cardioid directivity patterns have been considered.

3.3.2.3 Measuring the IVD Responses of a Real Microphone Array

After having analysed the effect of different microphone array characteristics on the

IVD measurements with simulated A-format microphone array, a method is devised to

measure the underlying space-frequency responses of a real microphone array in a re-

verberant environment. This is done to understand the behaviour of a real microphone

array with respect to IVD measurements. Another motivation behind such microphone

array analysis is to use the underlying microphone array response to perform location-

based correction of IVD measurements, as later proposed in Section 3.4.

The method to obtain the IVD response of a real microphone array is outlined in Fig.

3.11. The aim of the method is to measure the room impulse responses (RIR), and

later distinguish the microphone array responses from room responses. In this work,

microphone array was placed at the center of the room to avoid as much early reflections

as possible. It was mounted on an automatic turntable which was capable of rotating

in steps of 5 degrees to give accurate location based microphone array responses. The

Page 71: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 50

Sign

al

A-Format Sinesweep IR’s

Rotate the Microphone in Steps

0 100 200 300 400 500 600 700 800 900 1000-0.06

-0.04

-0.02

0

0.02

0.04

0.06

Microphone Location-Frequency Lookup Table

Delete Room Response Keep Mic Response

Frequency (Hz)

Figure 3.11: Method to obtain location-based correction coefficients

Page 72: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 51

loudspeaker was placed at a distance of 1.5m, and both the loudspeaker and the mi-

crophone array were at a height of 1.3m above the floor (see Fig. 3.10). At each step

of microphone array rotation, RIR was measured using the sine-sweep technique [124].

Due to rotation of the microphone array on its own axis, the room response remained

constant for each RIR measurement, while the microphone array response varied.

Subsequently, the obtained RIRs were analysed to find the arrival time of the first

reflections in order to get rid of the room responses. The RIRs were processed to keep

only the direct responses by deleting the room responses after approximately 3ms of

the arrival time of direct sound component.

The isolated direct sound components were zero padded before Fourier transform to

have a sufficient frequency resolution of IVD measurements. As a result, IVD measure-

ments were carried out to obtain location-based microphone array correction coefficients

using single frames containing direct responses at different rotations of the microphone

array. The obtained correction coefficients are depicted in the bottom part of Fig. 3.11

and also separately in Fig. 3.13 (the similarity with simulated A-format array IVD

measurements presented in Fig. 3.7c is noticeable). Due to the significance of time

of arrival of the first reflection, the recording arrangement and room geometry impact

on the effective frequency range of IVD correction coefficients. As a result, in our

experiments the lower limit of the effective range is approximately 330 Hz, which can

be improved by carrying out the measurements at a point where there are no nearby

reflective surfaces.

3.3.2.4 Effect of Microphone Array Response

The real microphone array response is analysed with respect to IVD measurements

in order to understand the behaviour and potential of the utilised microphone array.

The same concepts may be implied for any microphone array which uses pressure and

pressure gradient signals for intensity measurements. The measured space-frequency

response from 70 degree and 110 degree locations is shown in Fig. 3.12. With increase

in frequency, in Fig. 3.12 the deteriorating space-frequency response can be seen as

the deviation of IVD measurements from 70 degree and 110 degree source locations.

Page 73: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 52

Figure 3.12: Space-frequency microphone response for 70 degrees and 110 degrees. The

deteriorating space-frequency response is shown in this figure. It is represented by

the deviation of IVD measurements from the physical location of the sources. After a

certain limit frequency (approx. 9200 Hz for the array used) the IVD measurements of

both the locations can be seen to overlap in space. As this happens the source content

originating from these locations is no longer spatially independent. The response of the

array at 70 degrees can be seen to phase wrap around after the limit frequency due to

aliasing.

After a certain frequency the IVD measurements of both the locations can be seen to

overlap in space. As this happens the source content originating from these locations

is no longer spatially independent. It implies inadequacy of space-frequency sparsity in

a broader spectrum to perform source separation using IVD measurements [95][96].

The complete space-frequency response of the A-format microphone array is depicted

in Fig. 3.13 with location based IVD measurements every 5 degree interval. Four

distinguished groups can be identified as four quadrants, with space-frequency response

converging at 0 degrees, 90 degrees, 180 degrees and 270 degrees. The reason behind

better space-frequency response at these angles is the way cardioid microphones are

located to form a symmetrical structure as discussed earlier in Section 3.3.2.1.

Page 74: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 53

Figure 3.13: The figure depicts IVD measurements of the microphone array throughout

the space. It is the complete space-frequency response, with each line representing the

IVD measurements from a location in space. Again the space-frequency response is

seen to deteriorate significantly after the limit frequency (approx. 9200 Hz). Four

distinguished groups can be identified as the four quadrants, with space-frequency

response converging at 0 degrees, 90 degrees, 180 degrees and 270 degrees. After the

limit frequency is reached (shown as the point of convergence of IVDs) the spatial

independence of different locations is compromised due to aliasing at high frequencies.

3.3.3 Effect of the Room Environment

The second factor analysed in this section with respect to IVD-based system is the room

environment transfer function. The space-frequency room responses in the form of IVD

measurements for 70 degree and 110 degree source locations are shown in Fig. 3.14.

If a comparison is made with the microphone array response given in Fig. 3.12, the

overall trend is the same; however, due to reflections and reverberations there is now a

spread of IVD measurements around the source locations. It is a result of interaction

of each time-frequency component with the surrounding room environment.

In a room environment, the impact of reflections and reverberations is dependent on

several factors including the room geometry, location of acoustic sources, location of

Page 75: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 54

2000 4000 6000 8000 10000 12000 14000 16000

0

20

40

60

80

100

120

140

160

180

Frequency (Hz)

An

gle

(D

eg

ree

s)

Location = 70 DegreesLocation = 110 Degrees

Figure 3.14: Space-frequency response of ITU-R BS1116 standard listening room with

T60 of 0.32s at 70 degree and 110 degree locations. The room response is shown to

result in spread of IVD measurements around the physical location of the sources.

microphone array and reflective index of the room. To understand their impact on

source separation system, analysis based on IVD measurements (equation 3.10) and

acoustic energy (E) (equation 2.31) has been carried out to visualise the energy distri-

bution in space. For each STFT frame the energy distribution in surrounding space is

measured as:

E (θ; (ω, t)) =

360∑IV D=1

E (IV D; (ω, t)) . (3.15)

Four acoustic sources, each two and a half seconds in length, have been utilised for

depicting the energy distribution in space. These include English male speech, English

female speech, cello music and trumpet music. These sources are taken from an ane-

choic recordings dataset recorded as part of a European project named Archimedes

[125][130]. They have been low pass filtered with 8000 Hz cut off frequency, to get rid

of significantly distorted space-frequency response based on analysis presented earlier

in Section 3.3.2.4.

Page 76: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 55

50 100 150 200 250 300 3500

500

1000

1500

2000

Angle (Degrees)

En

erg

y

(a) Trumpet Music at 40 degree

50 100 150 200 250 300 3500

1000

2000

3000

4000

Angle (Degrees)

En

erg

y

(b) Oldgavotte Music at 130 degree

50 100 150 200 250 300 3500

500

1000

1500

2000

2500

3000

Angle (Degrees)

En

erg

y

(c) Female Speech at 220 degree

50 100 150 200 250 300 3500

500

1000

1500

2000

2500

3000

3500

Angle (Degrees)

En

erg

y

(d) Male Speech at 310 degree

Figure 3.15: Energy distribution in space of four sources using the microphone array

response only. The energy distribution is observed to be very compact in space. This

figure exhibits the potential of IVD-based system with the microphone array responses

low pass filtered at 8000Hz for single sources located at 40, 130, 220 and 310 degrees.

While using only the microphone array responses, Fig. 3.15 demonstrates the IVD and

E measurement of four sources placed one by one at 40, 130, 220 and 310 degrees. The

IVD gives the direction of the time-frequency component in space while E depicts their

energy. The measurements been shown collectively for all the processed STFT frames.

As shown in Fig. 3.15, when alone, each source has a very compact spatial signature,

i.e. the perceived location of the source through IVD and energy measurements is quite

precise in space.

The results with impulse responses including the effect of surrounding room environ-

ment are shown in Fig. 3.16. The spatial signature of each source through IVD and

Page 77: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 56

50 100 150 200 250 300 3500

200

400

600

800

1000

1200

1400

Angle (Degrees)

En

erg

y

(a) Trumpet Music at 40 degree

50 100 150 200 250 300 3500

500

1000

1500

2000

Angle (Degrees)

En

erg

y

(b) Old Gavotte Music at 130 degree

50 100 150 200 250 300 3500

1000

2000

3000

4000

5000

Angle (Degrees)

En

erg

y

(c) Female Speech at 220 degree

50 100 150 200 250 300 3500

500

1000

1500

Angle (Degrees)

En

erg

y

(d) Male Speech at 310 degree

Figure 3.16: Energy distribution of four sources in ITU-R BS1116 standard listening

room with T60 of 0.32s. The energy distributions of the sources in space are observed

to spread around the physical locations of the sources (40, 130, 220 and 310 degrees)

as a result of the effect of room response. The energy distributions without the room

response have been depicted in Fig. 3.15.

energy measurements is observed to spread around the source locations. From the liter-

ature it is found that for close to anechoic environment, IVD modelled with Laplacian

distribution perform source modelling better, but for more reverberant environments

the von-Mises distribution is preferred [131]. It is observed here that the spread of

source energy in space would require appropriate beam-width of the von-Mises spatial

filter in order to capture the desired source signal appropriately. As such, the spatial

filter should be capable of adapting to such energy spread around the desired source

location while using a non-exhaustive, computationally light technique.

Page 78: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 57

0 100 200 3000

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

DeterminedTrue

(a) Original: 0.32s Reverberation Time

0 100 200 3000

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

DeterminedTrue

(b) Corrected: 0.67s Reverberation Time

Figure 3.17: Mean and Standard Deviation of IVD measurements with different rever-

beration times - 300Hz to 8000Hz

To analyse the effect of location-based room responses on IVD measurements, the IVD

statistics of two rooms with 0.32s and 0.67s reverberation times (T60), using 36 circular

measurements (every 10 degree angular interval) were evaluated. The circular mean

and standard deviation of the IVD measurements for both the rooms are pictorially

represented in Fig. 3.17. The room with a greater reverberation time is observed

to have a greater standard deviation of IVD measurements. In the room with 0.67s

reverberation time, presence of equipment such as large display screens and furniture

resulted in the varying reflective index at different source locations, because of which

the spread varies at different angular intervals. As such, strong reflections reduce the

DRR and greatly impact on the IVD measurements. Numerically, the average IVD

statistics of both rooms are presented in Table 3.1.

Table 3.1: IVD measurement statistics for T60 = 0.32s and 0.67s (300Hz - 8000Hz)

T60 = 0.32s T60 = 0.67s

Standard Deviation 13.44◦ 17.12◦

Mean 5.95◦ 6.36◦

This analysis shows that the spread of IVD measurements in space is dependent on the

Page 79: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.3. Analysis of the IVD-based System 58

location of sources and the level of reverberation in the surrounding room environment.

Again, such observation complements the requirement of adaptive spatial filters that

adapts according to the locally isolated IVD statistics, while not knowing the location-

based room responses beforehand.

3.3.4 Interaction Amongst Sources

Although the IVD-based source separation system performs on the assumption that

sources are W-disjoint orthogonal (see Section 3.2.2) [11], when multiple sources are

active, similar to other sparsity-based algorithms, the IVD-based algorithm also suffers

due to violation of this condition [132][65]. In this section the energy distribution with

single source scenarios is compared with the energy distribution of a four sources sce-

nario (Fig. 3.18) to demonstrate the impact of multiple sources on IVD measurements.

For four sources being active simultaneously, the source locations are kept the same as

those of single source scenarios (Fig. 3.15) in order to make a fair observation.

In Fig. 3.18 the energy distribution of multiple sources scenario can be seen in com-

parison to single source measurements given in Fig. 3.15. The spatial signature of

each source is again observed to have expanded. In this case however, only due to the

presence of other active sources. The peak energy of each source has also diminished,

indicating the spread of energy in space.

The analysis presented in Section 3.3.1, Section 3.3.3 and Section 3.3.4 complements

the requirement of a source separation system capable of adaptation to the varying

mixing conditions. Theoretically the system should be able to blindly cater for the

effect of angular interval between sources, the varying spectral content of the sources

and their location-based interaction with the surrounding room environment.

Based on the analysis of IVD-based source separation system three modifications,

termed as microphone array correction, cardioid weight matrix, and adaptive spatial

filters are proposed in the sections that follow.

Page 80: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 59

50 100 150 200 250 300 350

0

500

1000

1500

2000

2500

3000

3500

Angle (Degrees)

En

erg

y

Figure 3.18: Energy distribution of four simultaneously active sources at 40, 130, 220

and 310 degrees, considering only the microphone array responses, has been shown

to depict the comparison with energy distributions of single sources active alone (Fig.

3.15). The effect of overlapping time-frequency content among sources results in the

spread of energy distributions in space as compared to compact energy distribution of

sources active alone, as shown in Fig. 3.15.

3.4 Microphone Array Correction

Farina [133] and Faller [95] have proposed solutions to deal with deviation of micro-

phone array response from ideal behaviour. Farina’s method is measurement-based

[133], but it only corrects microphone response along the x, y and z axis. Faller’s ap-

proach however is purely theoretical. The Minimum Mean Square Error (MMSE) based

approach proposed by Faller does not give suitable correction filters for real microphone

arrays [95].

Here a measurement-based approach applicable to the IVD-based source separation

system is proposed which consumes less processing time. The location-based space-

frequency responses of the microphone array measured off-line are used as a look-

Page 81: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 60

up table to perform correction of the IVD measurements throughout the space. The

method for acquiring the location-based look-up table, containing microphone array

correction coefficients, is presented in detail in Section 3.3.2.3. The location-based

space-frequency correction coefficients are shown in Fig. 3.13 after every 5 degree

angular interval.

DFT

IVDMeasurement

A-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

ŝn (t)

Spatial Filters

IDFT

Source Locations

W(ω,t)

Microphone Array

Correction

Figure 3.19: Block diagram of the proposed IVD-based microphone array correction

algorithm. The IVD measurements are corrected using the look-up table and source

locations before applying the spatial filters.

For desired source locations, the nearest space-frequency correction coefficient vector

is chosen and IVD measurements are corrected using equation 3.16. Subtraction and

resetting of the IVD measurements is carried out to achieve correction as:

Cn(ω, t) = IV D(ω, t)− Zn(ω, t) + (Ln ×Q). (3.16)

Here Cn(ω, t) represents the corrected IVD measurements for n-th source location using

the nearest space-frequency correction coefficient vector, Zn(ω, t). The index of the

chosen correction vector is Ln, and Q is the resolution factor of correction. Index of

the correction vector is the index of location-based response of the microphone array

which is closest to the desired source location. The resolution factor depends on the

impulse response measurements made to identify the microphone array response from

Page 82: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 61

Figure 3.20: Corrected space-frequency response of the microphone array shown in

intervals of 10 degrees. The response is observed to improve below the limit frequency

as compared to original IVD measurements depicted in Fig. 3.13. See Table 3.2,

Table 3.3, Table 3.4 and Table 3.5 for the statistics of corrected IVD measurements in

comparison with the original measurements.

different locations in space. The corrected space-frequency response after every 10

degree interval is shown in Fig. 3.20.

3.4.1 Evaluation of Microphone Array Correction

In this subsection the evaluation of proposed microphone array correction algorithm is

presented for one speaker scenarios in two different rooms. In the first part of evaluation,

microphone array responses are separated from room responses for both the rooms. As

no reverberation is considered, in this part of evaluation these are mentioned as Room

1 and Room 2. The correction is performed at every 10 degree interval in terms of

varying physical location of the source in the surrounding space. The localisation

performance with and without location-based microphone correction is evaluated for

comparison purposes. The evaluation has been carried out utilising the mean and

standard deviation of IVD measurements considering RIRs for time-frequency units

Page 83: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 62

lying between 300Hz and 8000Hz.

0 100 200 3000

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

DeterminedTrue

(a) Original Microphone IVD statistics

100 200 300

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

CorrectedTrue

(b) Corrected Microphone IVD statistics

50 100 150 200 250 300 350−20

−15

−10

−5

0

5

10

15

20

Angle(Degrees)

Err

or (

Deg

rees

)

(c) Original Microphone IVD error statistics

50 100 150 200 250 300 350−15

−10

−5

0

5

10

15

Angle(Degrees)

Err

or (

Deg

rees

)

(d) Corrected Microphone IVD error statistics

Figure 3.21: Mean and Standard Deviation of original (red) and corrected (green) IVD

measurements - 300Hz to 8000Hz. See text for discussion.

The comparison of the IVD measurements with and without location-based correction

is depicted in Fig. 3.21. The error bar denotes the standard deviation of IVD mea-

surements around estimated source locations. The original IVD measurements can be

observed to follow a zigzag pattern around the true linear variation of source locations

(see Fig. 3.21a and Fig. 3.21c), converging at specific points in space. It is due to

the arrangement of cardioid microphones to form a nearly coincident array (see Sec-

tion 3.3.2.1 for details). The location-based corrected IVD measurements, depicted in

Page 84: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 63

Fig. 3.21b and Fig. 3.21d, provide a much better localisation of the time-frequency

components by utilising the correction look-up table. Numerically, the overall localisa-

tion error and the standard deviation of the IVD measurements considering 36 circular

locations for both the rooms is given in Table 3.2 and Table 3.3. The corrected mean

localisation error is reduced by more than half of the original localisation error.

Table 3.2: Microphone Correction Statistics Room 1 (300Hz - 8000Hz)

Original Corrected

Standard Deviation 3.85◦ 1.21◦

Mean 5.72◦ 2.94◦

Table 3.3: Microphone Correction Statistics Room 2 (300Hz - 8000Hz)

Original Corrected

Standard Deviation 3.69◦ 1.15◦

Mean 5.63◦ 2.75◦

In the second part, to evaluate the microphone array correction for one source scenarios

in a reverberant environment, evaluation of microphone array correction algorithm is

done with room responses. The evaluation is carried out for two rooms with 0.32s and

0.67s reverberation times (T60) using their RIRs. The IVD statistics for the original

and corrected room responses are pictorially represented in Fig. 3.22. The increase

in localisation error after the inclusion of room responses is noticeable when compared

with Fig. 3.21. The room with a greater reverberation time is observed to have a greater

standard deviation of IVD measurements. In the room with 0.67s reverberation time,

the presence of equipment such as large display screens and furniture resulted in the

varying reflective index at varying source locations. As such, strong reflections reduce

the DRR and greatly impact on the IVD measurements. The average of the determined

and location-based corrected IVD statistics are presented in Table 3.4 and Table 3.5.

The mean localisation error in both the rooms is reduced to approximately half of the

original localisation error.

Page 85: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.4. Microphone Array Correction 64

The proposed correction technique can also be used for room correction by including

the complete room responses while constructing the look-up table. It will lead to a

much better localisation and separation performance; however, as a result the system

will become room specific. It is important to note that due to a transfer function

mismatch, using the correction coefficients for a different room can actually degrade

the system performance.

0 100 200 3000

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

DeterminedTrue

(a) Original: 0.32s Reverberation Time

100 200 300

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

CorrectedTrue

(b) Corrected: 0.32s Reverberation Time

0 100 200 3000

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

DeterminedTrue

(c) Original: 0.67s Reverberation Time

100 200 300

50

100

150

200

250

300

350

Angle (Degrees)

An

gle

(D

eg

ree

s)

CorrectedTrue

(d) Corrected: 0.67s Reverberation Time

Figure 3.22: Original and Corrected Mean and Standard Deviation of IVD with different

reverberation times - 300Hz to 8000Hz

Page 86: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.5. Cardioid Weight Matrix 65

Table 3.4: Microphone Correction Statistics 0.32s Reverberation (300Hz - 8000Hz)

Original Corrected

Standard Deviation 13.44◦ 12.84◦

Mean 5.95◦ 2.79◦

Table 3.5: Microphone Correction Statistics 0.67s Reverberation (300Hz - 8000Hz)

Original Corrected

Standard Deviation 17.12◦ 16.53◦

Mean 6.36◦ 3.32◦

3.5 Cardioid Weight Matrix

DFT

IVDMeasurement

A-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

Spatial Filters

IDFT

Source Locations

ŝn (t)

Cardioid Weight Matrix

Figure 3.23: Block diagram of the proposed system to weight the optimum A-format

microphones using source locations.

The cardioid microphones generating pressures signals PLFU , PRFD, PLBD and PRBU

are oriented towards 45 degree, 315 degree, 135 degree and 225 degree in the horizontal

plane respectively, represented here as θLFU , θRFD, θLBD, and θRBU . For each source

location, the two closest cardioid signals are identified by angular measurement, termed

Page 87: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.5. Cardioid Weight Matrix 66

Figure 3.24: A-format microphone top view. The closest cardioid microphones are

identified for each source location (highlighted in green). Based on the angular inter-

val, weighting of cardioid microphones is proposed to attenuate unwanted interference

and overlapping time-frequency components impinging at the microphone array from

opposite direction to that of the desired source.

θd. It was observed that the way these microphones are oriented in space, for any source

location only one or two cardioid microphones will be within 90 degree angular interval.

To help in visualising this, the top view of an A-format microphone array is depicted

in Fig. 3.24. The closest found cardioid microphones with respect to depicted source

location are coloured in green.

Cardioid weight matrix U is calculated based on the angular interval measurement,

θd, of each cardioid microphone from the given source locations. This cardioid weight

matrix has dimensions of (4×N), where N is the number of sources to be separated.

The weights for closest one or two A-format cardioid signals carry values depending on

their angular interval to source location, and the remaining cardioid signals are given

zero weighting to suppress unwanted directional information. As such, for each source

there are one or two non-zero values in its respective column vector, represented as

Page 88: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 67

Uan and U bn. The closer the source location to any of the oriented cardioid, the higher

the corresponding weight will be. The calculations are given in equation 3.17. Uan and

U bn are found using the two smallest angular intervals θan and θbn, among the angular

intervals of four cardioid microphones. The total contribution from the selected cardioid

signals is unity, Uan + U bn = 1. In case of a source exactly present towards any of the

four cardioid microphones, the respective Uan is 1 and U bn is zero.

(θan, θ

bn

)=

µn

θLFU

θRFD | θd < 90.

θLBD

θRBU

(3.17)

Uan = 1− θan90. (3.18)

U bn = 1− θbn90. (3.19)

The cardioid weight matrix algorithm greatly reduces the impact of other sources and

room reflections contributing to the mixture from other than the desired source loca-

tion. The significance of performance improvement in objective measures using cardioid

weight matrix will be presented later in Section 3.8.

3.6 Adaptive Spatial Filters

As a real world system is not an ideal theoretical system, by analysis presented in

Section 3.3 it is understood that the IVD measurements do not depend only on the

physical location of the sources. The measurement depends on the transfer function

of microphone array, transfer function of room environment and impact of multiple

sources on the desired source. As a result, the acoustic sources may be modelled with

von-Mises distribution function in real environments [11][131].

Page 89: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 68

DFT

IVDA-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

AdaptiveFilters

IDFT

Source Locations

W(ω,t)

ŝn (t)

Energy

Regions of Interest

Figure 3.25: Block diagram of the proposed system. The Energy and IVD measure-

ments are selected based on their closeness to desired source location using regions

of interest. Once the local statistics of the source are locally isolated, spatial filters

parameters are set accordingly.

To cater for the effect of varying angular interval between sources, varying content

of the sources, and the location-based interaction of the varying source content with

surrounding environment, adaptation of the spatial filter parameters is proposed in this

section. The spatial filters should be capable of reducing the beam-width if any nearby

source is active, also the spatial filter should be informed of the energy distribution

around desired source location for adapting the beam-width accordingly. As a result,

if the angular interval is less the spatial filter beam-width will reduce; however, for a

greater angular interval the spatial filter will adapt to optimum beam-width based on

energy distribution of the desired source.

For adaptation of the von-Mises spatial filters, f (θ;µn, σn), regions of interest, Øn, for

each source location µn are proposed. The regions of interest help in spatially isolating

the desired source in the surrounding space where one or more interfering sources could

be active. In these regions of interest we propose adaptation of the von-Mises spatial

filter parameters based on IVD and energy measurements.

Øn is set as an ideal circular binary filter, having the value of 1 at γ degrees at either

Page 90: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 69

50 100 150 200 250 300 350

0

0.2

0.4

0.6

0.8

1

Angle (Degrees)

Re

gio

n o

f In

tere

st

20 Degree70 Degree130 Degree200 Degree

Figure 3.26: Regions of interest for four source locations. The region of interest is an

ideal circular binary filter to locally isolate the IVD measurements of desired sources.

side of the desired source locations, and 0 otherwise:

f (Øn) =

1 µn − γ < Øn < µn + γ

0 otherwise.

(3.20)

The regions of interest provide ‘spatial windows’ to help exploit the local IVD statistics

of each source independently. Some regions of interests for different locations with γ

equal to 20 are depicted in Fig. 3.26.

From analysis presented in Section 3.3 it is understood that different aspects of the

source separation system, including surrounding room environment and the effect of

other active sources, impact on the IVD measurement of each time-frequency unit. As

a result the ‘focal point’ of each source shifts from its actual physical location over

short periods of time. To track shifts in the source ‘focal points’ and capture the

energy spread with optimum beam-width, IVD-based system needs to adapt. For each

region of interest the energy weighted mean and energy weighted standard deviation

are measured as:

Jn (t) = {IV D (ω, t) | IV D (ω, t) ∈ {Øn = 1}} (3.21)

En (t) = {E (ω, t) | IV D (ω, t) ∈ {Øn = 1}} (3.22)

Page 91: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 70

〈En (t)〉 =En (t)∑En (t)

, (3.23)

where Jn (t) and En (t) are a subset of IV D (ω, t) and E (ω, t) respectively for relevant

regions of interest. Energy is normalised to have the weights in between the range of

0 and 1. Energy weighted mean, µn(t), and standard deviation, σn(t), are calculated

following the spatial selection based on normalised energy. The angles are converted

into vectors rn(t) in two dimensional plane to facilitate the calculation of circular mean

and standard deviation. The vectors are averaged over frequency components to get a

resultant vector for every region of interest for the time frame t as [134]:

rn (t) =

∑rn (t)

length (Jn (t)). (3.24)

Here rn (t) is transformed using the four quadrant inverse tangent function to yield

the circular mean direction µn (t) for each source. The length of mean resultant vector

rn (t) indicates the circular spread [134]. The closer it is to one, the more concentrated

the energy distribution around the circular mean direction. It can be used analogous

to the linear standard deviation as in [134]:

Rn ( t) = ‖rn (t)‖ (3.25)

σn (t) = V√−2 lnRn (t), (3.26)

where σn (t) is the obtained standard deviation of source n at time frame t and V is

a scaling constant. Based on these two parameters, the spatial filters f(θ;µn(t), σn(t))

are updated for each STFT frame. To make this process robust to noise, smoothing is

carried out using a first order IIR filter. It was chosen due to its well known simplicity

and effectiveness [135].

As discussed in the analysis presented earlier in Section 3.2, at less angular intervals a

low spatial filter beam-width provides best performance. As such, to cater for physically

close sources, minimization of overlap between spatial filters to a certain threshold is

Page 92: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 71

proposed as part of the adaptive spatial filter algorithm. It is done by reducing σn(t)

of the analysed spatial filters found to have exceeded the spatial overlap threshold.

As later shown in Section 3.8, the proposed adaptive spatial filter algorithm greatly

reduces the interference coming in from other sources at less angular intervals. The

complete algorithm for adaptive spatial filtering is given in the pseudo code form in

algorithm 1.

Algorithm 1 Energy Weighted Adaptive Filter

1: function Adaptive Filter(Øn,µ(n,t−1),σ(n,t))

2: σ(n,t) = σ(n,t) ∗ α1 + σ(n,t−1) ∗ (1− α1)

3: µ(n,t) = µ(n,t) ∗ α2 + µ(n,t−1) ∗ (1− α2)

4: function Spatial Filter(σ(n,t), µ(n,t))

5: return f(θ;µ(n,t), σ(n,t))

6: end function

7: function Spatial Overlap(f(θ;µ(n,t), σ(n,t)))

8: if Spatial Overlap > S then

9: while Spatial Overlap > S do

10: µ(n,t) = µ(n,t−1)

11: σ(n,t) = σ(n,t) − SpatialOverlap

12: function Spatial Filter(σ(n,t), µ(n,t))

13: return f(θ;µ(n,t), σ(n,t))

14: end function

15: end while

16: else

17: return f(θ;µ(n,t), σ(n,t))

18: end if

19: end function

20: return f(θ;µ(n,t), σ(n,t))

21: end function

Page 93: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 72

3.6.1 Rate of Adaptation of Spatial Filter Parameters

The beam-width and location of the spatial filter in the respective region of interest is

updated based on the observed IVD statistics. To make this process robust to noise,

first order IIR filtering of the spatial filter parameters is carried out. In this section,

in order to understand the system behaviour, the parameters are evaluated using the

BSS Eval objective measures for different rates of adaptation. The rate of adaptation

of spatial filter location is denoted by α1 and that of filter beam-width is denoted by

α2. The average performance results of different combinations of two source mixtures,

varying from a minimum angle of 20 degrees to a maximum angle of 180 degrees, are

presented. Also the four source combinations are discussed, varying from a minimum

angular interval of 30 degrees to a maximum of 90 degrees. Adaptation of the spatial

filter parameters is done by varying α1 and α2 from 0.05 to 0.95 in steps of 0.1. A

higher value of the rate of adaptation denotes swiftly changing spatial filter parameter.

The two source SDR and SIR results for the varying rates of adaptation of spatial

filter parameters are presented in Fig. 3.27a and 3.27b respectively. The overall trend

of SDR and SIR measures is observed to be the same. The change in SIR measure,

however is not as significant as compared to the SDR. It is interesting to note that

the location and beam-width of the spatial filter behave inversely with respect to the

rate of adaptation. Swift changes in the spatial filter location using energy weighting

are observed to be significantly better than slowly changing location. It is reflected by

high values of the SDR and SIR when the adaptation rate of spatial filter location is

high (see Fig. 3.27a and Fig. 3.27b). The ability of a spatial filter to capture the ‘focal

point’ of the source, varying over short time durations, contributes to such improved

performance.

Page 94: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.6. Adaptive Spatial Filters 73

00.5

1.0

0

0.5

1

10

10.5

11

11.5

α1

α2

SD

R (

dB

)

10.2

10.4

10.6

10.8

11

11.2

(a) Two source SDR

00.20.40.60.81

0

0.5

1

21.4

21.6

21.8

22

22.2

22.4

α1

α2

SIR

(d

B)

21.6

21.7

21.8

21.9

22

22.1

(b) Two source SIR

00.20.40.60.81

0

0.5

1

5

5.5

6

6.5

α1

α2

SD

R (

dB

)

5.4

5.6

5.8

6

6.2

(c) Four source SDR

00.5

1

0

0.5

1

14.2

14.4

14.6

14.8

15

15.2

α1

α2

SIR

(d

B)

14.4

14.5

14.6

14.7

14.8

14.9

(d) Four Source SIR

Figure 3.27: Objective evaluation: Adaptation rate of spatial filter parameters

On the contrary, a high adaptation rate of spatial filter beam-width leads to poor

results. The overall output quality is observed to degrade as α2 increases. As the

energy weighted standard deviation directly relates to the adaptive spatial filter beam-

width, quick changes in the filter beam-width cause the spatial filter area to change

drastically. Such swift changes in the spatial filter area vary the amount of captured

source energy from one time frame to another, causing the objective output quality

to degrade. One apparent solution can be power normalisation of the spatial filter;

however, it will also lead to varying output sound energy due to the variation in spatial

filter magnitude.

The four source objective results are presented in Fig. 3.27c and Fig. 3.27d. In com-

parison to the two source scenario, the rate of change of spatial filter mean gives poor

Page 95: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.7. Combining All Modifications 74

results at very high adaptation rate. This is due to the presence of temporally less

consistent, more noisy IVD measurement with the increase in the number of sources.

The added reflections and reduced time-frequency sparsity among source signals, lead

to more dispersed IVD measurements in space. As such, a much lower adaptation

rate of the filter mean, than the two sources scenario, gives more reliable source esti-

mates. Similar to the two source scenario, with increase in rate of adaptation of filter

beam-width the overall quality of the output decreases. The SIR measure does not

change significantly throughout the varying rates of adaptation, except for the region

representing very low rate of adaptation of the spatial filter location.

To conclude, the spatial filter locations should be updated over short time durations

with IIR filtering corresponding to an averaging of approximately 50ms of audio data.

The spatial filter beam-width, however should be slowly updated with an IIR filtering

coefficient corresponding to around 500ms of audio data. If the number of sources is

given, then the adaptation rate can be set accordingly to have the optimum output

quality using the adaptive spatial filters. The source detection will be discussed in

detail in Chapter 4 where an algorithm for detecting the number of active sources is

proposed.

3.7 Combining All Modifications

Based on the same principle as proposed in the baseline IVD-based system [11], source

separation is finally achieved by soft masking IV D (ω, t) with the help of source specific

spatial filters, f(θ;µn(t), σn(t)). The difference due to the proposed modifications is

that the spatial filters are adaptively updated at every time frame to the objectively

evaluated best parameters, and instead of WB omnidirectional signal, the weighted A-

format cardioid signals are utilised to suppress unwanted directional information. The

source estimate sn (ω, t) are obtained as:

sn (ω, t) = Af (ω, t)× U × f(θ;µn(t), σn(t)). (3.27)

Here Af (ω, t) is the(

2M × 4)

A-format data matrix, which is weighed by cardioid

weight matrix, U . For N sound sources N von-Mises spatial filters are defined. At the

Page 96: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.7. Combining All Modifications 75

DFT

IVDA-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

AdaptiveFilters

IDFT

Source Locations

ŝn (t)

Energy

Regions of Interest

Microphone Array

Correction

Cardioid Weight Matrix

Figure 3.28: Block diagram of the proposed IVD-based source separation system (con-

tributions highlighted in green). Initially, the pressure and pressure gradient signals

are obtained from the microphone array. Then, the DFT of the signals are calculated.

Next, the intensity vector directions are calculated and consequently the IVD measure-

ments are corrected using the microphone array look-up table. Next, energy weighted

mean and standard deviation are used to adapt spatial filter parameters in regions of

interest that locally isolate IVD measurements of each source. Based on known source

locations, the cardioid weight matrix is determined. For the desired sources, adapted

von-Mises spatial filters are used to perform separation using appropriately weighed mi-

crophone signals based on cardioid weight matrix. Finally, the IDFTs of the separated

signals are calculated to get time-domain estimates of the sources.

final stage, after different source components have been identified and filtered within

each STFT frame, inverse discrete Fourier transform is calculated to transform sepa-

rated sources back to the time domain.

With all modifications combined into a single source separation system pipeline, the

proposed system pipeline is depicted in Fig. 3.28.

Page 97: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 76

3.8 Experimental Results

In this section first the computational cost of the proposed IVD-based source separation

system is empirically evaluated using computational time. All three modifications are

tested for mixtures of two and four sources for various source and location combinations

using BSS Eval toolbox [126]. The evaluation has been carried out in two rooms with

different reverberation times. In the end, the objective performance measures with the

proposed algorithms and the baseline IVD-based algorithm are compared.

3.8.1 Empirical Computational Time

It is important to analyse the performance of the proposed IVD-based source separation

algorithm to assess its low processing time claim. The performance of the proposed

algorithms can be measured empirically by assessing computational time of the BSS

demonstration developed in Matlab 2012b. A 2.3GHz and 6Gb RAM Intel machine

with profiling set to a single processor was used for this task. The built-in clock function

of Matlab 2012b was used for these measurements. The time taken to complete each

processing stage for every STFT frame was computed. The average time consumed

to process all STFT frames was taken as the time consumed to perform a specific

processing stage of the proposed system pipeline. In Fig. 3.29 the results are given

separately for the performance of upto four sources for repetitive steps, which are

carried out for each STFT frame, and non-repetitive steps, which are carried out once

at initialization.

Among the repetitive processing steps, the most time consuming is the adaptive filtering

algorithm. As the number of sources increases, the adaptive filter algorithm has greater

number of spatial filter pairs to check for correlation. The evaluated computational time

for the rest of the proposed processing steps does not show similar increase with increase

in number of sources. Among the initialisation parameters, calculation of regions of

interest shows increase in computation time with increase in the number of sources due

to their computation carried out in a programming loop.

Page 98: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 77

1 2 3 40

1

2

3

4 x 10−3

Number of Sources

Tim

e (

Se

co

nd

s)

DFTB−format ConversionAzimuth & EnergyCardioid SelectionAdaptive FilterMic CorrectionIDFT & BeamformingTotal Time

(a) Steps carried out for each time frame

1 2 3 40

0.5

1

1.5

2

2.5

3 x 10−3

Number of Sources

Tim

e (

Se

co

nd

s)

Mic Selection

Region of InterestCardioid Weight

Total Time

(b) Steps carried out at initialisation

Figure 3.29: Computational time of different processing steps in the proposed system

pipeline

The STFT frame size was 2048 samples with 50 percent overlap between adjacent

frames. At a sampling rate of 44100Hz this implies 23.2ms of algorithmic delay, i.e.

the system waits for the next 23.2ms of audio data to arrive at the input. The com-

putational time measurements of all the processing steps are demonstrating systems

potential to output the separated sound in much lesser time. This provides with addi-

tional time which can be utilised to plot real-time data for the BSS demonstration.

3.8.2 Two Source Mixtures

For two sources experiments, the first source was moved from 10 degree to 90 degree

anti-clockwise with 10 degree increments while the other was moved from 350 degree to

270 degree clockwise. In this manner the angular interval between two sources increased

from 20 degree up to 180 degree. The experiments were carried out in a similar manner

for both the rooms. The average for all the six two-source combinations at a particular

angular interval has been shown as the performance of the system. For comparison

with the baseline IVD-based source separation system, the best suitable beam-width

was chosen empirically by compiling results over all the source combinations at different

angular intervals.

The SDR and SIR results shown in Fig. 3.30 and Fig. 3.31 respectively, depict the

Page 99: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 78

20 40 60 80 100 120 140 160 180

2

4

6

8

10

12

14

16

Angular Interval

SD

R (

dB

)

(a) Microphone array correction

20 40 60 80 100 120 140 160 180

2

4

6

8

10

12

14

16

Angular IntervalSD

R (

dB

)

(b) Adaptive spatial filters

20 40 60 80 100 120 140 160 180

2

4

6

8

10

12

14

16

Angular Interval

SD

R (

dB

)

(c) Cardioid weight matrix

20 40 60 80 100 120 140 160 180

2

4

6

8

10

12

14

16

Angular Interval

SD

R (

dB

)

(d) Combining all modifications

Figure 3.30: SDR comparison baseline IVD system performance (red) with (a) Mi-

crophone array correction (b) Adaptive spatial filters (c) Cardioid weight matrix (d)

Combining all modifications, for two source scenario.

Page 100: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 79

baseline IVD-based system (red) in comparison to the proposed modifications (blue)

for two-source combinations. The microphone correction results, depicted in Fig. 3.30a

and Fig 3.31a, show nominal improvement in both the SDR and SIR measures. It is to

be noted that with the proposed microphone array correction algorithm, a significant

improvement is expected at much lesser reverberation time because the spread of the

IVD measurements is expected to be less. In such a situation the observed response will

be closer to the inherent microphone array response. Also the proposed microphone

array correction algorithm is more effective with single source localisation as the spread

of IVD measurements increases with added sources as well. Section 3.4.1 gives a more

detailed account of microphone correction for single source scenarios with different

reverberation times.

The SDR and SIR results with adaptive spatial filter implementation (Fig. 3.30b and

Fig. 3.31b respectively) show significant improvement at less angular intervals due to

minimization of the correlation between the spatial filters, and less significant at greater

angular intervals. On the other hand, the cardioid weight matrix algorithm, depicted in

Fig. 3.30c and 3.31c, was found to be the most effective in performance improvement

at greater angular intervals. The cardioid weight matrix only considers the data of

relevant cardioid microphones for separating the source of interest, as compared to the

omnidirectional response used by Gunel et al. [11]. The reflections and interference

contributing to the mixture from other than desired source location are completely at-

tenuated using the proposed cardioid weight matrix. The effect of non-ideal coincidence

of the cardioid microphones is also minimised with this modification as the nearest two

cardioid microphones are selected for weighting.

Finally the complete proposed system pipeline is shown to be notably superior at both

the lower and greater angular intervals when compared with the baseline system in

Fig. 3.30d and Fig. 3.31d. Similarly, the proposed system is shown to give significantly

better performance for a greater reverberation time in Fig. 3.32. An SDR and SIR

improvement of 1.96 dB and 2.96 dB respectively is reported with an ITU-R BS1116

standard listening room (T60 = 0.32s), and of 1.91 dB and 3.04 dB respectively with a

greater reverberation time (T60 = 0.67s).

Page 101: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 80

20 40 60 80 100 120 140 160 180

5

10

15

20

25

Angular Interval

SIR

(d

B)

(a) Microphone array correction

20 40 60 80 100 120 140 160 180

5

10

15

20

25

Angular IntervalSIR

(d

B)

(b) Adaptive spatial filters

20 40 60 80 100 120 140 160 180

5

10

15

20

25

Angular Interval

SIR

(d

B)

(c) Cardioid weight matrix

20 40 60 80 100 120 140 160 180

5

10

15

20

25

Angular Interval

SIR

(d

B)

(d) Combining all modifications

Figure 3.31: SIR Comparison of baseline IVD system performance (red) with (a) Mi-

crophone array correction (b) Adaptive spatial filters (c) Cardioid weight matrix (d)

Combining all modifications, for two source scenario.

Page 102: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 81

20 40 60 80 100 120 140 160 180

2

4

6

8

10

12

14

16

Angular Interval

SD

R (

dB

)

(a) SDR Comparison

20 40 60 80 100 120 140 160 180

5

10

15

20

25

Angular Interval

SIR

(d

B)

(b) SIR Comparison

Figure 3.32: Comparison of baseline IVD system performance (red) with all modifica-

tions combined (blue), for two source scenario, RT = 0.67s

3.8.3 Four Source Mixtures

For analysing the system performance with four source mixtures, four different scenarios

were designed as shown in Fig. 3.33. With respect to the positive x-axis, one source was

moved with 10 degree and the second with 30 degree clockwise steps. The other two

were moved in a similar manner but in anticlockwise steps. As a result, the smallest

angular interval between any two sources varied from 30 degree, 50 degree and 70

degree. For the final scenario the sources were placed along the four Cartesian axis

to give an angular interval of 90 degree between any two adjacent sources. The SDR

and SIR results are plotted in Fig. 3.34 and Fig. 3.35 respectively. With a greater

number of sources, the frequency overlap among source signals increases, and because

of added reflections, the spread of IVD measurements is greater. Hence overall system

performance degrades as compared to two source scenario.

It may be suggested that without the microphone array correction there is an unsym-

metrical bias in the IVD measurements with increase in frequency (see Fig. 3.14).

Using correction, the bias is reduced making the mean location of the IVD measure-

ments closer to the source location, facilitating the adaptive filter algorithm to capture

the varying IVD measurements around the source location more effectively.

Page 103: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 82

x

y

x

y

x

y

x

y

20⁰

50⁰

340⁰

310⁰

30⁰

80⁰

330⁰

280⁰

40⁰

110⁰

320⁰

250⁰

0⁰

90⁰

180⁰

270⁰

Figure 3.33: Different four sources scenarios considered for experiments - Minimum

angle between any two speakers varying from 30 degree to 90 degree in 20 degree

intervals to check systems performance for varying angular intervals.

Similar to the two sources scenario, the proposed adaptive filter algorithm is shown to

be performing in comparison to the best empirically evaluated results without the user

input or exhaustively searching for optimum beam-width values. With all the improve-

ments a significant increase in overall output sound quality is achieved as compared to

the baseline system.

The results for vision lab are plotted in Fig. 3.36, which also clearly show the superior

performance of the proposed system as compared to the baseline system. A noticeable

SDR and SIR improvement of 1.23 dB and 3.38 dB respectively with four sources is

reported for an ITU-R BS1116 standard listening room (T60 = 0.32 s) and, 1.21 dB

and 3.28 dB respectively with a greater reverberation time (T60 = 0.67 s).

Page 104: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 83

30 50 70 90

2

4

6

8

Angular Interval

SD

R (

dB

)

(a) Microphone array correction

30 50 70 90

2

4

6

8

Angular IntervalSD

R (

dB

)

(b) Adaptive spatial filters

30 50 70 90

2

4

6

8

10

Angular Interval

SD

R (

dB

)

(c) Cardioid weight matrix

30 50 70 90

2

4

6

8

10

Angular Interval

SD

R (

dB

)

(d) Combining all modifications

Figure 3.34: SDR Comparison of baseline IVD system performance (red) with (a)

Microphone array correction (b) Adaptive spatial filters (c) Cardioid weight matrix (d)

Combining all modifications, for four source scenario.

Page 105: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.8. Experimental Results 84

30 50 70 90

5

10

15

20

Angular Interval

SIR

(d

B)

(a) Microphone array correction

30 50 70 90

5

10

15

20

Angular IntervalSIR

(d

B)

(b) Adaptive spatial filters

30 50 70 90

5

10

15

20

Angular Interval

SIR

(d

B)

(c) Cardioid weight matrix

30 50 70 90

5

10

15

20

Angular Interval

SIR

(d

B)

(d) Combining all modifications

Figure 3.35: SIR Comparison of empirically best selected baseline performance (red)

(a) Microphone array correction (b) Adaptive spatial filters (c) Cardioid weight matrix

(d) Combining all modifications, for four source scenario.

Page 106: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.9. Conclusion 85

30 50 70 90

3

4

5

6

7

8

9

Angular Interval

SD

R (

dB

)

(a) SDR Comparison

30 50 70 90

5

10

15

20

Angular Interval

SIR

(d

B)

(b) SIR Comparison

Figure 3.36: Comparison of baseline IVD system performance (red) with all modifica-

tions combined (blue), for four source scenario, RT = 0.67s

3.9 Conclusion

The proposed algorithms to separate individual sound sources from the observed sound

mixtures have been shown to perform significantly better and more efficiently than the

baseline IVD-based system in real-time. While using proposed algorithms, the user can

listen to the active sources in desired combinations in real-time. A measurement-based

algorithm for microphone array correction is proposed to improve the separation perfor-

mance below limit frequency; however, the improvement is found to be less significant

with more reverberant room responses and greater number of sources. To cater for the

interaction of source with room environment and with other sources in a non-exhaustive

manner, energy weighted adaptive spatial filters are proposed to dynamically adjust to

optimum parameters. The cardioid weight matrix algorithm greatly reduces the impact

of other sources and room reflections contributing to the mixture from other than the

desired source locations. The adaptive spatial filters efficiently adjust to suitable filter

parameters at low angular intervals giving superior results, while at greater angular

intervals the cardioid weight matrix algorithm gives significantly better results. Com-

bining all modifications resulted in significant improvement at all the angular intervals.

Experiments with numerical evaluation were conducted to assess the performance of

Page 107: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

3.9. Conclusion 86

proposed algorithms over the baseline IVD-based source separation system. An average

SDR improvement of 1.96 dB with two sources and 1.23 dB with four sources is achieved

for an ITU-R BS1116 standard listening room (T60 = 0.32s). An improvement of 1.91

dB for two sources and 1.21 dB for four sources in SDR is achieved for a room with a

greater reverberation time (T60 = 0.67s). Similarly average SIR improvement of 2.96

dB with two sources and 3.38 dB with four sources is achieved for an ITU-R BS1116

standard listening room (T60 = 0.32s). An improvement of 3.04 dB for two sources and

3.28 dB for four sources in SIR is achieved for a room with a greater reverberation time

(T60 = 0.67s).

Page 108: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Chapter 4

Blind Source Separation of

Moving Speakers Based on

Intensity Vector Statistics

4.1 Preview

This chapter presents a novel approach to detect, track and separate moving speakers

in real-time using intensity vector direction (IVD) statistics. The aim is to update un-

mixing system parameters swiftly and accurately in order to deal with the time-variant

mixing parameters. Investigation of IVD distribution in space for practical acoustic sce-

narios is presented, which establishes a baseline for the proposed blind moving source

separation (BMSS) system. Denoising is carried out to extract reliable speaker esti-

mates using von-Mises modelling of the IVD measurements in space and IIR filtering of

the IVD distribution in time. Peaks in the IVD distribution are assigned location expec-

tations to check for consistency, and consequently high location expectation peaks are

declared as active speakers. Also, the proposed location expectation algorithm caters

for natural pauses during speech delivery. Speaker movements are tracked by spatial

isolation of the detected peaks using time-variant regions of interest. As a result, the

proposed system is capable of blind speaker detection, tracking and separation.

87

Page 109: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.2. Introduction 88

4.2 Introduction

Over the last decade considerable research has been carried out in the context of physi-

cally stationary sources [19][13][20][21][22], while moving sources have been notably less

investigated [23][18][24]. Movement of speakers makes the source separation problem

more challenging due to time-variant mixing parameters [62]. Unmixing parameters

cannot be accurately and swiftly measured due to lack of sufficient data over which

the sources are physically stationary [60]. In real world scenarios, the BMSS problem

is multifaceted [83]. It includes swift and accurate approximation of the room impulse

response (RIR), speaker activity detection (in case a speaker enters or exits the acous-

tic scene), tracking speaker movements with reliability, microphone array configuration

and the computational cost of the algorithm.

The BMSS problem for M microphones and N sources can be mathematically repre-

sented as:

X(t) = H(t)S(t). (4.1)

Here N source signals are represented as S(t) = [s1(t)s2(t)s3(t)...sN (t)]T while the

transfer function coefficients are contained in (M × N) mixing matrix H(t), such

that hmn(t) represents the coefficient for m-th microphone and n-th source at time

t. X(t) = [x1(t)x2(t)x3(t)...xM (t)]T contains the resultant mixture components of the

M microphones. The task of the BMSS system is to extract source signal S(t) from the

observed mixtures X(t). For a stationary source scenario the dependency of transfer

function, H, on time, t, is absent due to time-invariant transfer functions.

In practical scenarios, separation of moving acoustic sources is a very useful prepro-

cessing stage for speaker identification and speech recognition problems [62][136]: for

example, it is normal for a speaker to move while delivering a presentation. Several

other applications are also directly related to the BMSS system. These include noise

suppression, high quality hearing aids, acoustical surveillance, automatic camera con-

trol, 3D audio signal processing and forensic science for time-variant mixing conditions

[137][11][138].

Page 110: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 89

In this chapter as a proof of concept, a robust, low complexity and deterministic inten-

sity vector direction based BMSS system is presented. To the best of our knowledge,

no source separation system has been proposed for convolutive mixtures capable of de-

tection, tracking and separation of up-to four moving speakers. Based on the proposed

system pipeline a demonstration has been developed in Matlab 2012b (see Appendix

A), which facilitates user interaction with the surrounding moving speakers.

Section 4.3 presents analysis of the baseline IVD-based system (Section 3.2) with respect

to IVD distribution in space. It is presented to examine and understand the behaviour

and potential of IVD-based system in order to perform BMSS. Section 4.4 explains

the methodology of the proposed system with implementation details, and evaluates

the proposed parameters. Section 4.5 describes the experimental test conditions and

presents the obtained results, and Section 4.6 concludes the chapter.

4.3 Analysis of the IVD Distribution in Space

The distribution of IVD measurements in space is investigated in this section to un-

derstand the potential of IVD-based source separation system to deal with realistic

acoustic scenarios. In realistic acoustic scenarios, the blind source separation problem

is challenging due to practical constraints [139]. In such scenarios, the system has to

deal with dynamically changing acoustic environment, for example moving speakers or

time varying number of speakers. In order to effectively perform BMSS, the system

should have the ability to detect time-variant number of speakers, and track speaker

movements. The movements can be side to side (influencing the spatial signature of

the speakers) or backwards and forwards (determining the captured energy of speakers)

with respect to the microphone array.

The baseline IVD-based system (Section 3.2) is incapable of dealing with practical

acoustic scenarios because of its inability to perform speaker detection and tracking.

It is critical to investigate the IVD-based system with different acoustic scenarios due

to varying system behaviour with time-variant mixing conditions. In this section some

relevant observations are presented and discussed. All observations have been made

Page 111: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 90

with convolutive recordings of human subjects in an ITU-R BS1116 standard listening

room having a reverberation time of 320ms.

Speech signals are known to be quasi-stationary [140], which are defined as non-

stationary signals with locally stationary statistics over short periods of time that are

different from statistics of other local times [140]. As a result, STFT processing yields

valuable representation of speech signals in the time-frequency domain. This processing

gives representation of the speech data at scalar frequencies throughout the available

spectrum over short periods of time. It is dependent on the sampling frequency and

window size in samples. Although a large time window provides more frequency res-

olution and enables increased separability between multiple speakers, signals tend to

become non-stationary. Due to approximately stationary characteristics of speech sig-

nals in between 20ms and 40ms, a window size close to this time period is preferred

for speech related applications [141][142]. If the window size is reduced, the spectral

estimates tend to become unreliable due to ineffective frequency resolution and the

stochastic nature of speech over very short time durations. Another important aspect

to consider with much smaller window sizes is that the frame shift becomes smaller,

resulting in increased rate of frames being received and processed. As a result more

information is processed than required, increasing the computational cost and reducing

the algorithmic delay. For the system under investigation, a 46.4ms time window was

chosen as a trade off between these factors. With 44100 Hz sampling frequency, it is

representative of 2048 data samples in time. As such, the analysis of baseline IVD-

based system has been carried out with the Short Time Fourier Transform (STFT)

overlap and add (OLA) technique with 2048 window size and 50% overlap.

In the literature it is found that previously Laplacian and von-Mises distributions have

been used to model sparse data in space [11][143][110]. In this contribution, we model

the IVD in each time-frequency unit of the STFT with von-Mises distribution function

in order to deal with reverberant mixtures [11]. The von-Mises distribution function is

given as [144]:

f (θ; IV D (ω, t) , σ) =eσ cos(θ−IV D(ω,t))

2πIo (σ), (4.2)

Page 112: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 91

where IV D (ω, t) is the measured direction of a particular time-frequency unit, σ is

the concentration parameter, similar to the variance of Gaussian distribution, and θ is

the circular angle lying within 0 and 2π. The modified Bessel function of zero order is

represented by Io (σ), where σ is logarithmically related to the 6 dB beam-width θBW

as [11]:

σ =ln 2

1− cos(θBW2

) . (4.3)

θBW is chosen as a small value to maintain the localisation accuracy. For each pro-

cessed STFT frame, using the von-Mises distribution function, IVD distribution in the

surrounding space is measured as:

H (IV D; (ω, t)) =

360∑IV D=1

f (θ; IV D (ω, t) , σ) , (4.4)

where f (θ; IV D (ω, t) , σ) represents the IVD of a time-frequency unit modelled by the

von-Mises distribution function, as given in equation 4.2.

4.3.1 Acoustic Source Activity

When one or more speakers become active in the surrounding space, their activity is

reflected in the IVD distribution. The IVD distribution of two physically stationary

speakers is shown in Fig. 4.2 over a period of 2.5s after various processing steps.

The x, y and z axis represent the time frames, the surrounding space and the IVD

distribution respectively. The speakers are active at 45 and 135 degrees with respect

to the microphone array.

The raw IVD distribution (Fig. 4.2a) can be seen to have noisy information of the

two speakers. The frequency dependent environment transfer function, along with the

microphone array response contribute to the noisy time-frequency direction of arrival

(DOA) measurements [65]. Speech signals by nature are sparse in frequency and im-

pulsive in time [145], as a result the time-frequency components corresponding to the

Page 113: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 92

xα Σ

1-α z¯¹

y

+

+

Figure 4.1: First order IIR filter structure

same speaker may spatially be located away from each other by a small angular inter-

val. In addition, the unused time-frequency components behave like random noise due

to imperfections and undesirable components in the acquired data [123].

To smooth out noise and obtain reliable speaker information across time, we suggest

the use of a first order IIR filter due to its well known simplicity and effectiveness [135].

The implemented structure is depicted in Fig. 4.1. While processing from one STFT

frame to another, IIR filtering of the IVD distribution is carried out as:

H (IV D; (ω, t)) = H (IV D; (ω, t)) × α + H (IV D; (ω, t− 1)) × (1− α) , (4.5)

where IVD distribution of the current time frame, H (IV D; (ω, t)), is updated utilising

the IVD distribution of previous frame, H (IV D; (ω, t− 1) , ) based on the IIR filter

coefficient, α.

As a result of IIR filtering, depicted in Fig. 4.2b, a more distinguished and prominent

IVD distribution of each speaker is achieved. Similar to the first order IIR filtering,

which caters for noise in the time domain, denoising in space is achieved by mod-

elling each IVD using the von-Mises distribution function. The von-Mises modelling of

the IVD allows for a more noise attenuated speaker IVD distribution across space by

smoothing out the effect of noisy IVD measurements, as depicted in Fig.4.2c.

The overall impact of noise reduction across both time and space can be observed in

Fig. 4.2d. Such denoising results in a more robust IVD distribution representation of

speakers, which can be exploited to detect and track speaker movements over short

time durations with reliability.

Page 114: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 93

2040

6080100 50

100150

0

5

10

Angle (Degrees)

Time (Fra

mes)

IVD

Dis

trib

utio

n

0

5

10Speaker 1

Speaker 2

(a)

2040

6080100 50

100150

0

1

2

3

Angle (Degrees)

Time (Fra

mes)

IVD

Dis

trib

utio

n

1

2

3Speaker 1

Speaker 2

(b)

2040

6080100 50

100150

2

4

6

Angle (Degrees)

Time (Fra

mes)

IVD

Dis

trib

utio

n

1

2

3

4

5

6

Speaker 1Speaker 2

(c)

2040

6080100 50

100150

1

2

Angle (Degrees)

Time (Fra

mes)

IVD

Dis

trib

utio

n

0.5

1

1.5

2

2.5

Speaker 1Speaker 2

(d)

Figure 4.2: IVD Distribution in space (a) Without von-Mises modelling (b) Applying

first order IIR Filter to smooth out noise in the time domain (c) Modelling the IVD

with von-Mises distribution funtion to smooth out noise in space (d) Applying von-

Mises model along with first order IIR filter to achieve denoised IVD distribution in

both time and space.

Page 115: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 94

4.3.2 Effect of Distance to the Microphone Array

(a)

(b)

Figure 4.3: Effect of distance of the speaker to microphone array on IVD distribution

in space. (a) Ground truth (b) IVD distribution in space. Speaker 1 is moving forwards

and backwards w.r.t to the microphone array while Speaker 2 is stationary in space.

The closer the speaker gets, the greater its captured energy and its direct to reverberant

ratio. As a result it has better IVD distribution representation in space (prominent

with a high peak). The distribution of a physically stationary Speaker 2 is also shown

to get affected due to the change in distance of Speaker 1.

The distance of speaker to microphone array placed at the centre of the room dictates

two parameters: the speaker energy and its direct to reverberant ratio (DRR). Fig.

4.3 depicts the scenario of a speaker varying in distance to the microphone array while

another physically stationary speaker is active in the vicinity. Speaker 2 is at a constant

distance of 1m to the microphone array while Speaker 1 starts at 1m, comes as close

as 0.3m, and moves away to 1.6m, facing towards the microphone array all the time.

As Speaker 1 comes close to the microphone array, due to increase in recorded speaker

energy, comparatively low energy time-frequency components also start to contribute

Page 116: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.3. Analysis of the IVD Distribution in Space 95

to the speaker IVD distribution. It is reflected by a rise in the IVD distribution of

Speaker 1 (see Fig. 4.3). It is also to be noted that at the time frames when Speaker

1 is closest to the microphone array (Fig. 4.3: time frames 200 to 300), Speaker 2 is

observed to have the lowest IVD distribution.

As Speaker 1 moves away from the reflective walls and closer to microphone array, the

DRR gets higher as well. Consequently, a more prominent speaker IVD distribution

in space is achieved because of the reduced impact of reflections on the direct sound

components. As Speaker 1 starts moving away, the IVD distribution gets lower in space

(Fig. 4.3: time frame 300 onwards), and eventually Speaker 2 is more prominent when

the distance of Speaker 1 is greater than that of Speaker 2. As such, a distant low

energy speaker cannot be detected using IVD distribution, and unfavourably acts as

diffused background babble noise. Also if the microphone array is located near any

reflective objects, for example in the corner of a room, the speaker IVD distribution

gets dispersed in space, leading to less meaningful spatial signatures of speakers. Given

that both Speaker 1 and Speaker 2 have an approximately constant energy envelope, it

is concluded that the IVD distribution is dependent on relative energy of the speakers

as well as their absolute energy.

4.3.3 Moving Acoustic Sources

With speakers moving around in the surrounding environment, IVD-based system is

investigated to detect changes in the IVD distribution over short time durations in

order to perform speaker tracking. IVD distribution of two moving speakers over a

period of 16s is depicted in Fig. 4.4. The speaker movements can be seen to take place

in two distinct paths, highlighted within white boundary lines. The directivity of the

speakers was either sideways or towards the microphone array during their activity,

leading to limited variations in the captured energy and DRR.

Page 117: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 96

(a)

(b)

Figure 4.4: IVD distribution of two moving speakers at the same distance to the mi-

crohone array. (a) Ground truth (b) IVD distribution in space. Speaker 1 is moving

between 75 degrees and 125 degrees, while Speaker 2 is moving between 175 and 225

degrees.

In Fig. 4.4, the consistency of a speaker in space can be observed as it carries out its

movements. Considering such denoised observations of the time-variant IVD distribu-

tion in space, the speaker detection and tracking can be performed if the unmixing

parameters are estimated accurately and with reliability over short time durations. As

such, the IVD distribution can be exploited to perform source separation in practical

acoustic scenarios if the sources are comparable in energy levels. This leads us to the

proposed BMSS methodology, which exploits time-variant IVD distribution in space to

update the unmixing filters accordingly.

4.4 Proposed Methodology

The sparse time-frequency representation of acoustic signals can be achieved using their

W-disjoint orthogonality characteristic, which states that only one source is dominant at

Page 118: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 97

DFT

IVDMeasurement

A-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t) ŝn (t)

Spatial Filters

IDFT

Speaker Locations

Figure 4.5: Processing stages of the baseline IVD-based system [11] (see Section 3.2).

Initially, the pressure and pressure gradient signals are obtained from the microphone

array. Then, the DFT of the signals is measured. Next, the intensity vector directions

are calculated in time-frequency domain. Next, von-Mises spatial filters are applied for

the known speaker locations. Finally, IDFTs of the separated signals are calculated to

get the time-domain estimates of the target speakers.

a particular time-frequency unit [65]. Speech signals are known to be approximately W-

disjoint orthogonal in the STFT domain even in strong reverberations [110]. Therefore,

aiming to get sparse speaker representations, separation is carried out in the time-

frequency domain using STFT processing. The required pressure and pressure gradient

signals for the IVD measurements are obtained using a commercially available, nearly

coincident microphone array [118][122]. No prior information is assumed except the

configuration of the utilised microphone array, allowing the separation to take place in

a blind manner.

4.4.1 IVD Distribution in Space

In Section 4.3 some acoustic scenarios have been analysed with respect to IVD distri-

bution in space. Based on the analysis presented, peaks in IVD distribution correspond

to potential speaker activity. In case of multiple simultaneously active speakers, the

greater the number of time frames used for measuring the IVD distribution in space,

the greater is the confidence in speaker locations; however, this is no longer true for

Page 119: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 98

moving speakers. As a result, detecting speaker locations as well as having the ability

to perform tracking over short time durations is critical for the performance of BMSS

system.

There are several hurdles in exploiting the IVD distribution in space for performing

source separation in practical situations. As presented earlier in Section 4.3, IVD

distribution in space is not spatially compact as the actual physical location of the

sources, it is dependent on several factors which include reverberant response of the

surrounding room environment, relative energy of the speakers, their DRR, microphone

array response, the impact of multiple speakers on a particular time-frequency unit and

other imperfections in the acquired data [123][65].

4.4.1.1 Impact of the Frequency Range on IVD Distribution in Space

In Fig. 4.6 it is shown that due to sparse nature of speech signals [145], the IVD

measurements not dominated by any of the active speakers may lead to less accurate

speaker detection and tracking over short time durations. In Fig. 4.6 the effect is

visualised using the IVD distribution of two stationary speakers with varying frequency

range of the IVD measurements utilised for IVD distribution representation. The IVD

distribution is shown over a period of four seconds, with Speaker 2 being a distant low

energy speaker.

The utilisation of time-frequency units up to 4000Hz (4.6a) to up to 22050Hz (4.6d)

have been depicted. Speaker 2, which is distant (approx. 3 meters) to the microphone

array portrays less prominent IVD distribution with increase in the frequency range

until it completely becomes part of the IVD distribution floor. Excessively unused

time-frequency components lead to such an observation because of which a relevant

low frequency range for IVD distribution representation of speakers is preferred.

A much broader frequency range also leads to more spread out IVD distribution in space

due to poor space-frequency response of the microphone array at higher frequencies

(see section 3.3.2.1). In light of these observations, a lower frequency range, which is

spatially more compact and provides reliable speaker estimates, is preferred for IVD

distribution representation.

Page 120: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 99

(a) Frequency bins up till 4000 Hz (b) Frequency bins up till 8000 Hz

(c) Frequency bins up till 16000 Hz (d) Frequency bins up till 22050 Hz

Figure 4.6: The utilisation of time-frequency units up to 4000Hz to up to 22050Hz for

IVD distribution representation. Speaker 2, which is distant (approx. 3 meters) to

the microphone array portrays less prominent IVD distribution with increase in the

frequency range until it completely becomes part of the IVD distribution floor. Also

with the increase in frequency range, IVD distribution of Speaker 1 can be seen to

spread in space.

4.4.1.2 Effect of First Order IIR Filter

A first order IIR filter has been proposed earlier in Section 4.3 to reduce noise of the

IVD distribution as it varies from one time frame to another. In the context of BMSS

Page 121: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 100

system, noise implies noisy IVD measurements which lead to inaccurate and unreliable

estimates of time-variant mixing parameters over short time durations, such as the

number of speakers and their locations. The impact of IIR filter coefficient used to

reduce noise and achieve reliable speaker estimates is analysed in this section.

The nature of IIR filter allows IVD distribution to be averaged over a time interval

dependent on the filtering coefficient. The first order IIR filter coefficient α, given in

equation 4.5, is varied from 0.5, 0.125, 0.0833 to 0.0625 while considering the STFT

window size equal to 2048 samples with an overlap of 50% between adjacent frames

at 44100Hz sampling frequency. It represents an effective averaging time of 46.4ms,

185.8ms, 278.6ms and 371.5ms respectively.

The results of two moving speakers delivering 10s of speech are depicted in Fig. 4.7

with respect to log IVD distribution. The results of peak detection and location ex-

pectation algorithms are depicted in Fig. 4.8 only to show the effect of IIR filter

coefficient on the peaks in IVD distribution. A more detailed discussion on the peak

detection and location expectation algorithms will follow in Section 4.4.2 and Section

4.4.3 respectively.

From Fig. 4.7a to Fig. 4.7d, a less difference between the minimum and maximum

values of the log IVD distribution can be observed due to a longer averaging time.

With a low effective averaging, in Fig. 4.7a the speech signals can be clearly seen

to appear as bursts in the IVD distribution because of their varying intensity over

short time durations. Also, the speaker IVD distributions are more spread out and

less consistent in space (Fig. 4.7a). This leads to noisy measurements of the IVD

distribution peaks, as depicted in Fig. 4.8a. As the effective averaging time increases,

the IVD distribution becomes more regular and denoised to reveal reliable estimates of

the speakers (Fig. 4.7b and Fig. 4.8b); however, averaging over a much longer period

of time leads to poor speaker estimates, because longer averaging times may result in

multiple peaks for the same speaker due to a slow update rate of the IVD distribution

(Fig. 4.7c and Fig. 4.8c, and Fig. 4.7d and Fig. 4.8d).

Page 122: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 101

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160

−2

−1.5

−1

−0.5

0

0.5

Speaker 1

Speaker 2

(a) IIR Filtering with α = 0.5

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160−1

−0.5

0

0.5

Speaker 1

Speaker 2

(b) IIR Filtering with α = 0.125

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160 −1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

Speaker 1

Speaker 2

(c) IIR Filtering with α = 0.0833

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

Speaker 1

Speaker 2

(d) IIR Filtering with α = 0.0625

Figure 4.7: Log IVD distribution with IIR filtering corresponding to different averaging

times (a) 46.4 ms (b) 185.8 ms (c) 278.6 ms (d) 371.5 ms. As the effective averaging time

increases, the IVD distribution becomes more regular and denoised to reveal reliable

estimates of the speakers; however, averaging over a much longer period of time leads

to poor speaker estimates, because longer averaging times may result in multiple peaks

for the same speaker due to a slow update rate of the IVD distribution (4.8).

Page 123: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 102

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

(a) IIR Filtering with α = 0.5

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

(b) IIR Filtering with α = 0.125

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

(c) IIR Filtering with α = 0.0833

Time (Frames)

An

gle

(D

eg

ree

s)

50 100 150 200 250 300 350 400

20

40

60

80

100

120

140

160 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

(d) IIR Filtering with α = 0.0625

Figure 4.8: Peak Detection & Location Expectation with IIR filtering corresponding to

different averaging times (a) 46.4 ms (b) 185.8 ms (c) 278.6 ms (d) 371.5 ms. The red

cross indicates the time delay caused in peak detection due to longer averaging times.

See text for discussion.

In addition to achieving a denoised IVD distribution in space, added processing steps

are required such as peak detection in the IVD distribution and assigning location

expectations to the detected peaks in order to have reliable speaker estimates over

short time durations in dynamically changing mixing conditions.

Page 124: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 103

4.4.2 Peak Detection

As the number of speakers increase, the IVD distribution gets distorted due to greater

number of overlapping time-frequency components and added reflections, and it is

dependent on the frequency content of speakers [69]. The average IVD distributions

in space of two and four physically stationary speaker mixtures over 2.5s of activity

have been depicted in Fig. 4.9a and 4.9b respectively. As shown in Fig. 4.9b, with

lower peaks and a higher distribution floor, the significance of speaker IVD distributions

decrease with the increase in the number of speakers.

When the environment is silent, there is a uniform distribution of IVD measurements

in space. As two speakers becomes active, the speaker IVD distribution peaks are high

and prominent with a low IVD distribution floor as shown in Fig. 4.9a. As more

speakers start to contribute to the mixture, the reflections and reverberations increase.

Also the speakers start to overlap in the time-frequency domain due increased violation

of W-disjoint orthogonality [69]. As a result, when compared to a scenario with less

number of active speakers, the speaker IVD distribution peaks are less prominent and

the floor is higher (Fig. 4.9b). Due to less significant IVD distribution of the speakers

with increased number of speakers, when added speakers carry out movements, the

speaker detection, tracking and separation becomes a challenging problem to solve over

short time durations.

In this work a peak detection algorithm based on derivatives [146] is used to detect

peaks for the updated IVD distribution at each time frame, represented as:

θn = Peaks {H (IV D; (ω, t)) , c (t)} (4.6)

where Peaks {} represents the utilised derivative-based peak finding function, c (t)

represents the time varying threshold and θn represents the estimated peaks. The

importance of having a time-variant threshold, c (t), is discussed later on in this section.

It is to be noted that any peak detection algorithm capable of selecting peaks in circular

data beyond a defined threshold can be utilised.

Several steps are proposed here to consider a peak as a potential speaker, while having

the least number of false detections:

Page 125: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 104

100 200 3000

0.5

1

1.5

2

2.5

3

Angle (Degrees)

IVD

Dis

trib

utio

nAvg. Peak

Avg. Floor

(a)

100 200 3000

0.5

1

1.5

Angle (Degrees)

IVD

Dis

trib

utio

n

Avg. Peak

Avg. Floor

(b)

Figure 4.9: Average IVD distribution measurements over 2.5s of recorded speech ac-

tivity: (a) Two Speakers (b) Four Speakers. The average IVD distribution peaks are

higher and more prominent for less speakers. With a greater number of speakers the

IVD noise floor gets high due to added reverberations and increase in overlapping time-

frequency content of speakers, resulting in a poor IVD distribution representation. Due

to poor IVD representation over short time durations, added speakers pose a challenge

to the moving blind source separation system.

1. At every time frame the IVD distribution floor is approximated by averaging the

determined troughs. The average distribution floor is termed K (t). It is used

later as an offset for peak detection.

2. Each detected peak is allocated a region of interest as an ideal circular binary

filter spanning γ degrees on its either side to allow for interference free movement

and satisfactory source separation. We chose γ equal to 25 degrees based on

the the evaluation of baseline IVD-based system presented in Section 3.3, where

it was presented that the sources had to be located sufficiently away from each

other to give satisfactory separation results. The chosen value leads to a possible

simultaneous definition of a maximum of 7 non-overlapping regions of interest.

3. With the increase in the number of speakers it is assumed that the speaker IVD

distribution peaks degrade linearly. The value beyond which the peak is consid-

ered a potential speaker is dynamically set as:

Page 126: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 105

c (t) = − 1

QN (t) + K (t) + 1. (4.7)

Here c (t) represents the threshold for peak detection, which is varying with the

number of speakers, N (t), and the time-variant estimated IVD distribution floor,

K (t). Q is set to 7 based on the maximum number of possible speakers. In this

manner the effective threshold for peak detection is dynamically lowered in order

to perform reliable speaker detection with added number of speakers.

4. If any of the found θn is closer than γ degrees, the peak with the higher IVD

distribution is considered while the other is ignored because a lower IVD distri-

bution peak indicates a weak speaker, and as such the more prominent speaker

is chosen.

After the IVD distribution peaks have been identified for potential speaker activity at

a given time frame, the IVD distribution may change due to change in the content

of speaker or speaker movements in the time frames that follow. As a result, to keep

track of the detected peaks, time-variant regions of interest, Øn(t), are defined. Similar

to Øn, as proposed in the previous chapter in Section 3.6, the time-variant region of

interests Øn(t) are set as ideal circular binary filters, having a value of 1 at γ degrees

on either side of detected speaker peak, θn, and 0 otherwise.

If the peak relocates in a particular region of interest due to speaker movement, it is

linked to the previously detected peak. The definition of time-variant region of interest

facilitates in locally isolating the speaker IVD distribution peak to perform tracking.

As a result, a potential speaker can be tracked independent of other speakers. Each

region of interest is updated in an adaptive manner at every time frame based on the

movement of the IVD distribution peak within it.

4.4.2.1 The Impact of Adaptive Peak Detection Threshold

The scenario depicted from Fig. 4.10a to Fig. 4.10c, representing the average IVD dis-

tribution over 2.5s of two, four and six speakers activity, shows the degrading speaker

Page 127: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 106

representation with the increase in number of speakers. The red line depicts the ap-

proximated average IVD distribution floor whereas the blue line depicts the average

of IVD distribution peaks. The difference between the average floor and average peak

is greater for the two speaker scenario. A high average peak value implies that the

threshold for peak detection should be kept high for reliable speaker detection with

the least possible false detections; however, with increasing number of speakers, the

threshold has to be lowered to carry out effective detection.

The detection performance with a high static threshold for different number of speaker

mixtures is shown from Fig. 4.10d to Fig. 4.10f. As shown in figure Fig. 4.10d, a

high static threshold allows for the two speaker mixture IVD distribution peaks to be

detected with accuracy; however, for the four speaker mixture, the detection is not

accurate for more than half the time of their activity (Fig. 4.10e), and six speakers

are not detected at all (Fig. 4.10f). It is to be noted that at times when a speaker

is absent for a short period of time and another is active, the IVD distribution of the

active source will be prominent, allowing it to be detected by the system.

A contrary scenario with a low static threshold used for peak detection is depicted

from Fig. 4.10j to Fig. 4.10l. In this scenario many false detections take place which

are highlighted with red ellipses. The proposed approach based on adaptation of the

threshold dependent on the number of speakers and the approximated IVD distribution

floor is depicted from Fig. 4.10g to Fig. 4.10i. It provides reliable detection results

with more accurate speaker detection rates as compared to static thresholding.

The system decreases the threshold value linearly as more speakers are detected to be

part of the mixture. It is very unlikely that all four or all six sources become active at

exactly the same time frame. This allows for the IVD based system to ‘change gears’

for detecting more speakers that may join in the acoustic scene.

Page 128: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 107

100 200 3000

1

2

3

Angle (Degrees)

IVD

Dis

trib

utio

n

Avg. Peak

Avg. Floor

(a) 2 Sources

100 200 3000

0.5

1

1.5

Angle (Degrees)

IVD

Dis

trib

utio

n

Avg. Peak

Avg. Floor

(b) 4 Sources

100 200 3000.2

0.4

0.6

0.8

1

Angle (Degrees)

IVD

Dis

trib

utio

n

Avg. Peak

Avg. Floor

(c) 6 Sources

500 1000 1500 20000

1

2

3

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(d) 2 Source - Detection

500 1000 1500 20000

2

4

6

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(e) 4 Source - Detection

500 1000 1500 20000

2

4

6

8

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(f) 6 Source - Detection

500 1000 1500 20000

1

2

3

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(g) 2 Source - Detection

500 1000 1500 20000

2

4

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(h) 4 Source - Detection

500 1000 1500 20000

2

4

6

8

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(i) 6 Source - Detection

500 1000 1500 20000

2

4

6

8

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(j) 2 Source - Detection

500 1000 1500 20000

2

4

6

8

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(k) 4 Source - Detection

500 1000 1500 20000

5

10

Time Frames

No

. o

f So

urc

es

DeterminedTrue

(l) 6 Source - Detection

Figure 4.10: (a)-(c) IVD distribution for different number of speakers (d)-(f) Peak

Detection results - high static threshold = 1.5 (g)-(i) Peak Detection results - adaptive

to number of speakers detected (j)-(l) Peak Detection results - low static threshold

= 0.5. (Red ellipses show errors. See Section 4.4.2.1 for discussion.)

Page 129: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 108

4.4.3 Location Expectation and Source Detection

Along with start-stop nature of speech within and in between words, a speaker may give

comparatively longer pauses in between sentences. It is important for user experience

that a speaker is tracked with these naturally varying pauses. Also an inactive speaker

should be labelled as absent to update the unmixing parameters of the acoustic scene

as swiftly as possible. For similar reasons, the noisy IVD distribution peaks need to

be catered for, as they represent random false speakers. To cater for such variations

in speaker activity and deal with noisy IVD distribution peaks over short time dura-

tions in order to perform reliable speaker detection and tracking, an expectation based

algorithm is proposed here.

After detecting the IVD distribution peaks as potential speakers and defining time-

variant regions of interest, the peaks are checked for consistency in the time frames

that follow. For each detected peak, a location expectation, εn, is assigned. The

process of assigning location expectations to the IVD distribution peaks is outlined

first, followed by a detailed discussion on the parameters.

The value of εn, defined to be bound between 0 and 1, is independent for each region

of interest, and updated as:

1. If a peak appears outside any already defined region of interest, a new region of

interest is defined. If a peak is detected in a particular region of interest, εn is

incremented by a predefined value p.

εn = εn + p; (4.8)

2. If no peak is detected in a particular region of interest, the corresponding εn is

decremented by a predefined value q.

εn = εn − q; (4.9)

3. If no peak is found then the value is decremented over several time frames to zero

and the corresponding region of interest is deleted.

Page 130: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 109

Consequently, the speaker detection is based on location expectation εn. The location

expectation value equal to 0.5 is chosen as the threshold for speaker detection, as it

provides equal scope for increment to 1 and decrement to 0. If the expectation is greater

than 0.5, it is labelled as an active speaker, otherwise not:

f (εn) =

1 εn ≥ 0.5

0 εn < 0.5,

(4.10)

where f (εn) equal to 1 means the speaker is declared as active. It is to be noted that if

a speaker moves outside its region of interest and becomes active again it is considered

as a new speaker due to the systems inability to perform classification.

4.4.3.1 Evaluation of Increment and Decrement in Location Expectation

based on Speaker Activity

The increment in location expectation is used to identify more consistent IVD distri-

bution peaks because a consistent peak is a reliable indicator of an active speaker. A

slow increment in location expectation implies that the consistency of a peak is mon-

itored over a longer period of time, ensuring a reliable but delayed speaker detection.

The increment is varied from instantaneous speaker detection (Fig. 4.11a) to a delay

representing 0.075 s (Fig. 4.11b), 0.150 s (Fig. 4.11c) and 0.3 s (Fig. 4.11d) for four

moving speaker mixture. The false peaks are circled in red for clarity in analysis. The

amount of false detections can be observed to reduce as the delay in speaker detection

increases from an instantaneous value to 0.3s.

Page 131: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 110

Time Frames

An

gle

(D

eg

ree

s)

500 1000 1500 2000

50

100

150

200

250

300

350 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

Speaker 3 Speaker 4

(a) Instantaneous detection

Time Frames

An

gle

(D

eg

ree

s)

500 1000 1500 2000

50

100

150

200

250

300

350 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

Speaker 3 Speaker 4

(b) Increment representing delay of 0.075s

Time Frames

An

gle

(D

eg

ree

s)

500 1000 1500 2000

50

100

150

200

250

300

350 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

Speaker 3 Speaker 4

(c) Increment representing delay of 0.150s

Time Frames

An

gle

(D

eg

ree

s)

500 1000 1500 2000

50

100

150

200

250

300

350 0

0.2

0.4

0.6

0.8

1

Speaker 1

Speaker 2

Speaker 3 Speaker 4

(d) Increment representing delay of 0.300s

Figure 4.11: Peak detection and location expectation using different increment values

in the location expectation using p. The false peaks are circled in red for clarity in

analysis. The amount of false detections can be observed to reduce as the delay in

speaker detection increases from an instantaneous value to 0.3s.

Although a small increment in the location expectation gives reliable speaker detection

rates, the time delay in speaker detection is large. To depict the trade off, small

regions of time and space from Fig. 4.11 are shown in Fig. 4.12. The time frames

when a speaker becomes active have been depicted with varying increment values. As

shown with the help of red arrows, there is a greater delay in speaker detection with

a smaller increment in the location expectation, p, which increases from Fig. 4.12a to

Fig. 4.12d. Notice that in Fig. 4.12 a straight line of detection is shown for simplicity,

Page 132: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 111

Time Frames

An

gle

(D

eg

ree

s)

445 450 455 460 465 470 475 480

186

188

190

192

194

1960

0.2

0.4

0.6

0.8

1

(a) Instantaneous increment to 1

Time Frames

An

gle

(D

eg

ree

s)

445 450 455 460 465 470 475 480

186

188

190

192

194

1960

0.2

0.4

0.6

0.8

1

(b) Increment representing delay of 0.075s

Time Frames

An

gle

(D

eg

ree

s)

445 450 455 460 465 470 475 480

186

188

190

192

194

1960

0.2

0.4

0.6

0.8

1

(c) Increment representing delay of 0.150s

Time Frames

An

gle

(D

eg

ree

s)

445 450 455 460 465 470 475 480

186

188

190

192

194

1960

0.2

0.4

0.6

0.8

1

(d) Increment representing delay of 0.300s

Figure 4.12: Peak detection and location expectation using different increment values

in location expectation using p. Red arrows represent delay in time while detecting a

new speaker. See the text for discussion.

a moving peak is incremented in the location expectation value in a similar manner.

Similar to location expectation increment, p, which is used to perform reliable speaker

detection, the location expectation decrement, q, is utilised to cater for natural pauses

during speech activity and to label a no longer active speaker as absent. The lower

the value of decrement, the greater the time allowed for a speaker to pause in between

speech delivery. It was found that if the decrement value is too high, the speaker will

appear to switch on and off and will jump between output channels, whereas a low

value leads to an inactive speaker being labelled as active for a longer period of time.

Fig. 4.13 depicts the scenario of a speaker becoming quiet and then active again after

Page 133: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 112

about 4.5s. Fig. 4.13a to Fig. 4.13d present the increasing waiting time (labelled with

green arrows) for declaring the speaker as inactive. The location expectation falls below

the value of 0.5 in all the depicted scenarios, only the waiting time for the system to

label the speaker as inactive is shown to vary. By subjectively evaluating the acquired

moving speaker mixtures, it was found that a time duration of 1.2s is adequate in

declaring a speaker as completely inactive.

Time Frames

An

gle

(D

eg

ree

s)

1050 1100 1150 1200 1250

64

66

68

70

72

0

0.2

0.4

0.6

0.8

1

(a) Decrement representing wait of 0.4s

Time Frames

An

gle

(D

eg

ree

s)

1050 1100 1150 1200 1250

64

66

68

70

72

0

0.2

0.4

0.6

0.8

1

(b) Decrement representing wait of 0.8s

Time Frames

An

gle

(D

eg

ree

s)

1050 1100 1150 1200 1250

64

66

68

70

72

0

0.2

0.4

0.6

0.8

1

(c) Decrement representing wait of 1.2s

Time Frames

An

gle

(D

eg

ree

s)

1050 1100 1150 1200 1250

64

66

68

70

72

0

0.2

0.4

0.6

0.8

1

(d) Decrement representing wait of 2.4s

Figure 4.13: Peak detection and location expectation using different decrement values

in the location expectation by varying q. Green arrows show the waiting time before

declaring a speaker as absent. See the text for discussion.

Page 134: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.4. Proposed Methodology 113

DFT

IVDMeasurement

Location Expectation

A-format to

B-format

PLBD(t)

PRBU(t)PLFU(t)

PRFD(t)

IVDDistribution

z¯¹

Peak Detection

ŝn (t)

Spatial Filters

IDFT

Figure 4.14: Block Diagram of Proposed Moving BSS System (contributions highlighted

in blue). Initially, the pressure and pressure gradient signals are obtained from the

array. Then, the DFT of the signals are calculated. Next, the intensity vector directions

are calculated and consequently IIR filtered IVD distribution in space is measured using

von-Mises modelling of the IVD measurements. Next, peak detection is done to locate

potential speakers in the surrounding space. Location expectations are assigned to

the identified peaks for reliable speaker detection and to cater for pauses in between

speech activity. For the detected speakers time-variant von-Mises spatial filters are

designed and applied. Finally, IDFT of the separated signals are calculated to get the

time-domain estimates of target speakers.

4.4.4 Spatial Filtering

Finally, the spatial filters are formed with directivity towards locations which carry

a high expectation value. Source separation is achieved by soft masking the time-

frequency components to filter out IVD measurements with the help of source specific

spatial filters [11]. The source estimates sn(ω, t) are obtained as:

sn(ω, t) = WB(ω, t)f(θ;µ(n,t), σn

), (4.11)

where f(θ;µ(n,t), σn

)represents the von-Mises spatial filter having the time-variant

location µ(n,t) and static beam-width σn.

Page 135: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 114

At the final stage, after different source components have been identified and filtered,

inverse discrete Fourier transform is carried out to have the time domain source es-

timates. The designed BMSS system pipeline with all the modifications to baseline

IVD-based system is shown in Fig. 4.14. As a result of proposed modifications the

system is capable of dealing with dynamically changing acoustic scenes.

The detected speakers are displayed on the graphical user interface of BMSS demonstra-

tion (see Appendix A) facilitating the user to choose and listen to the desired speakers.

Note that the contributions proposed in Chapter 3 can be utilised to improve the out-

put quality of the BMSS system while exploiting the stationarity of speakers over short

time durations.

4.5 Objective Evaluation of the Proposed BMSS System

In this section the empirical computational cost of the proposed BMSS system is eval-

uated first. Then the recording set-up to perform objective evaluation of the system

is presented, followed by performance measures and the results of speaker detection,

tracking and separation. The evaluation of separated outputs is carried out using the

BSS Eval [126] and PEASS [147] toolboxes for the proposed BMSS algorithm while

using stationary and moving speaker scenarios, with two and four speaker mixtures.

4.5.1 BMSS Empirical Computational Time

Similar to the stationary blind source separation algorithm, the BMSS algorithm is

also claimed to be a low delay algorithm, which requires low processing time. The four

foundational steps of the BMSS system have been evaluated in Matlab 2012b using

a single 2.3 GHz CPU and 6 Gb RAM. The built-in clock function of the Matlab

2012b was utilised to carry out time measurements. The time taken to complete each

processing stage, for every STFT frame was computed. The average time taken to

process all the STFT frames was taken as the time consumed to perform a specific

processing stage of the proposed system pipeline. The results are depicted in Fig. 4.15.

Page 136: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 115

The algorithmic delay of the STFT processing was 23.2ms with 2048 sample window

size and 50% overlap between adjacent STFT frame. As can be seen in Fig. 4.15, the

first order IIR filter has the least computational cost due to its well known simplicity

[135]. The location expectation and peak detection algorithms follow, for which the

time taken is approximately constant with increasing number of speakers. The most

time consuming algorithm among the proposed is the IVD distribution measurement

due to its processing in a programming loop. Keeping in mind the algorithmic delay

of 23.2ms, none of the proposed algorithms are consuming too much time. As such

the BMSS algorithm starts to play the separated speakers from the current time frame

before the data of the next time frame arrives at input. The total time consumed by

these four processing steps with six speakers is less than 4ms on the tested platform.

2 3 4 5 60

1

2

3

4

x 10−3

Number of Sources

Tim

e (

Se

co

nd

s)

IVD HistogramIIR FilterPeak DetectionLocation Expectation

Figure 4.15: Computational time of proposed BMSS algorithms with respect to number

of speakers. See the text for discussion.

For other processing steps of baseline IVD-based implementation, the computational

costs are given in Section 3.8.1. Overall, the combined computational cost of the

proposed algorithms together with other fundamental processing steps such as the DFT,

A-format to B-format conversion, IVD measurement, spatial filtering and IDFT is less

than the algorithmic delay allowing the proposed algorithm to perform in real-time.

Page 137: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 116

4.5.2 Experimental Set-up

20

5 c

m

(a)

20

5 c

m

(b)

Figure 4.16: Experimental Setup: (a) Stationary source locations (b) Regions of speaker

movement. A maximum 60 degree interval of speaker movement is assigned for each

region in the designed set-up. It is to be noted that the proposed system can perform

speaker detection, tracking and separation even if the speaker moves out of the allocated

region, as long as it does not co-locate with other active speakers.

In order to obtain mixture recordings, speakers were asked to read pre-defined texts

while being stationary and while walking inside the ITU-R BS1116 standard listening

room (T60 = 0.32s). The movements were carried out in allocated regions to avoid

co-location in space. The stationary source locations and moving speaker regions are

depicted and labelled in Fig. 4.16a and Fig. 4.16b respectively. The recording time

of all the speakers varied between 20s and 30s. The speakers were directed to pause

approximately halfway into the recordings to test the BMSS system with varying mixing

parameters (increasing and decreasing number of speakers).

The sampling rate of the recordings was 44100Hz, made with Core Sound TetraMic

microphone system. The microphone array was set-up 1.6m high above the floor, close

Page 138: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 117

to average height of the speakers to demonstrate the horizontal BMSS system. The

recording point was at the center of the room. The acoustic axis (directivity) of the

speakers was towards or sideways upon the microphone array to capture direct sound

components, while the microphone array remained fixed. The distance of the speakers

to the microphone array was fixed to 1.2m. For stationary and moving speaker scenar-

ios, individual A-Format recordings were made. Three male speakers and one female

speaker participated in the recordings. The speaker mixtures were later obtained by

summing the reverberant recordings of the individual speakers in desired combinations.

This method exploits linearity and time-invariance of linear acoustics as carried out in

[11].

4.5.3 Objective Performance Measures

The BSS Eval [126] and PEASS [147] toolboxes were used to evaluate the performance

of stationary and moving speakers. The BSS Eval measures, Signal to distortion ratio

(SDR), signal to interference ratio (SIR) and signal to artefacts ratio (SAR) quantify

the separated output signals in terms of overall sound quality, interference and arte-

facts respectively. Using BSS Eval, the estimated source, si, of the true source, si, is

decomposed as [126]:

si = starget + einter + eartef + enoise. (4.12)

Here starget is the modified version of true source by an allowed distortion. einter, eartef

and enoise represent the interference, artefacts and noise error terms respectively. With

the help of such breakdown, the SDR is calculated as [126]:

SDR = 10 log 10Starget

2

(einter + enoise + eartef )2. (4.13)

At less angular interval between the speakers SIR measure gives reliable information by

quantifying the interference coming into the desired source from other active speakers.

The SIR is given as [126]:

SIR = 10 log 10Starget

2

einter2. (4.14)

Page 139: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 118

Similarly, independent of the interference and noise terms, the signal to artefacts ratio

(SAR) is measured as [126]:

SAR = 10 log 10(Starget + einter + enoise)

2

eartef 2. (4.15)

A more recent set of auditory-motivated objective measures based on improved dis-

tortion decomposition have been proposed in [147]. These measures include a set of

perceptually motivated objective measures that aim to predict the subjective scores

with accuracy [147]. These measures include source image to spatial distortion ratio

(ISR), SIR, SAR and SDR [147]. ISR represents the quality in terms of preservation of

target source as compared to overall distortion which is measured by SDR. Collectively

known as the Perceptual Evaluation methods for Audio Source Separation (PEASS)

[147], the new measures include target related perceptual score (TPS) and overall per-

ceptual score (OPS). TPS measures the closeness of separated signal with the clean

reference signal and OPS gives an overall quality score [147]. The reference signal

chosen for evaluation was the omnidirectional pressure signal.

4.5.4 Detection, Tracking and Separation - BMSS

The recordings were summed to acquire b+ d, b+ e, b+ f , b+ g and b+ h two speaker

mixtures for different regions labelled in Fig. 4.16. As examples, the source detection

and tracking for b+h and b+e mixtures is depicted from Fig. 4.17a to Fig. 4.17d. The

outcome of peak detection and the location expectation algorithms is depicted in Fig.

4.17a and Fig. 4.17c, while the respective time varying number of speakers detected

by the proposed system are shown in Fig. 4.17b and Fig. 4.17d. The true activity of

the speakers is shown in comparison with the activity determined by proposed BMSS

system.

In Fig. 4.17b and 4.17d, the speakers are silent in the beginning and are detected as

they become active. The dependency on time and location both is unknown, making

the system completely blind to the possible varying mixing parameters. As a speaker

becomes active, the peak detection and location expectation algorithms label it as

active after approximately 190ms of consistent peak detection.

Page 140: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 119

Time (Frames)

An

gle

(D

eg

ree

s)

500 1000 1500 2000

−180

−120

−60

0

60

120

0 0.5 1

(a) Speaker tracking b+ h

500 1000 1500 20000

1

2

3

Time (Frames)

No

. o

f So

urc

es

DeterminedTrue

(b) No. of sources detected b+ h

Time (Frames)

An

gle

(D

eg

ree

s)

500 1000 1500 2000

50

100

150

200

0 0.5 1

(c) Speaker tracking b+ e

500 1000 1500 20000

1

2

3

Time (Frames)

No

. o

f So

urc

es

DeterminedTrue

(d) No. of sources detected b+ e

Figure 4.17: Examples of two moving speakers - Tracking and detection performed

by the proposed BMSS system. In (a) and (c) the location expectation of speakers is

depicted to show tracking. Separate ‘tracks’ of speaker movement can be observed in

space. Near the time frames in the middle (800 - 1200), the speakers take a long pause

and start speaking again while carrying out movements. In (b) and (d) the true speaker

activity is compared with the activity determined by the proposed BMSS system.

Page 141: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 120

The tracking results in Fig. 4.17a and Fig. 4.17c depict the zigzag movement carried

out by speakers with respect to the microphone array. The two distinct tracking streams

and a fall in the detected number of speakers around the middle time frames can be

observed in the results given in Fig. 4.17. The continuity in speaker activity based on

time and space is broken around the middle time frames (800 to 1200) and speakers

are detected again with new mixing parameters.

In Fig. 4.17a and Fig. 4.17c, speaker location expectations can also be seen to slowly

decrease as they become quiet. The waiting time after the pause in activity around

the middle time frames is included as the true speaker time. It is to be noted that if a

speaker becomes quiet, moves outside the assigned region and becomes active again the

BMSS system will label it as a new speaker because of the proposed systems inability

to classify time-frequency components other than utilising their spatial consistency.

The objective evaluation of the proposed BMSS system is done for speakers labelled

as active, therefore such delay in speaker detection is penalised during the objective

evaluation. The PEASS and BSS Eval objective evaluation results for two moving

speaker separation are shown in Fig. 4.18 and Fig. 4.19 respectively. Four implemen-

tations have been shown in comparison, the baseline IVD-based system tested with

stationary sources is termed ‘Static Filter-Static Sources’, the baseline IVD-based sys-

tem tested with moving sources is termed ‘Static Filter-Moving Sources’, the proposed

BMSS system tested with stationary sources is termed as ‘BMSS-Static Sources’ and

the proposed BMSS system tested with moving sources is termed as ‘BMSS-Moving

Sources’. Both the PEASS and BSS Eval measures show fairly consistent results, with

the baseline IVD-based system performing worst for moving speakers because of its in-

ability to perform speaker tracking before carrying out separation. Comparatively, the

proposed BMSS algorithm gives better results by tracking the moving speakers before

performing source separation.

It is to be noted that the proposed system can perform speaker detection, tracking and

separation even if the speaker moves out of the allocated region, as long as the speaker

does not co-locate with other active speakers. Due to the designed experimental set-

up, results are biased in favour of the baseline IVD-based implementation; because the

Page 142: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 121

further a speaker moves from the static filter the worst its separation quality. In such a

scenario the proposed BMSS system will be able to perform speaker tracking, and give

more superior separation results than the baseline IVD-based system.

As the regions are further apart, the performance is shown to improve because of the

greater spatial discrimination, which leads to less interfering time-frequency compo-

nents in the desired source. For comparison purposes, separation for stationary sources

was evaluated with the BSS Eval and PEASS for two speakers using the proposed

BMSS system and the baseline IVD-based implementation to show the performance of

proposed system in stationary sources scenario. Overall the static sources were found to

have the best performance because of their stationary physical locations which lead to

time-invariant mixing parameters. The proposed BMSS system performs on par with

the IVD-based implementation, making it a suitable choice for dealing with stationary

speakers as well, with added capability of detecting and localising simultaneously active

sources.

It is important to note that the separation performance is dependent on relative energy

and DRR of the speakers as well. As shown in the analysis section (Sec. 4.3), with

increasing energy and DRR the impact of a speaker on time-frequency units becomes

greater, leading to better representation of speaker in the time-frequency domain. Sta-

tionary speakers have better DRR as compared to moving speakers due to their direc-

tivity towards the microphone array, as a result contributing to better objective results

of stationary sources.

For four speaker mixtures, the BMSS system was tested with a+ c+ e+ g and b+ d+

f + h speaker mixtures. As an example, the results of speaker tracking and detection

for a + c + e + g are depicted in Fig. 4.20a and 4.20b respectively. The BSS Eval

and PEASS performance results for the four speaker separation considering moving

and stationary sources are given in Fig. 4.21 and 4.22. The results show similar

trends but with overall decline in the separation quality as compared to two speaker

scenarios. The performance degrades with increased number of speakers due to added

reflections and reverberations in the surrounding environment. The added speakers

also lead to more overlapping time-frequency components, rendering the speaker IVD

Page 143: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 122

(a) (b)

(c) (d)

(e) (f)

0

10

20

30

40

50

bd be bf bg bh

OP

S

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

2

4

6

8

10

12

bd be bf bg bh

ISR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

5

10

15

20

bd be bf bg bh

SIR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

2

4

6

8

10

12

14

16

bd be bf bg bh

SAR

(d

B)

Static Filter-Moving SourcesBMSS-Moving SourcesBMSS-Static SourcesStaticFilter-StaticSources

0

2

4

6

8

10

bd be bf bg bh

SDR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

10

20

30

40

50

60

70

bd be bf bg bh

TPS

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

Figure 4.18: Two speaker mixtures: PEASS Toolbox results showing comparison be-

tween baseline IVD implementation and BMSS system with moving speakers and sta-

tionary speakers. See text for discussion.

Page 144: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 123

(a) (b)

(c)

0

2

4

6

8

10

bd be bf bg bh

SDR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

5

10

15

20

25

bd be bf bg bh

SIR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

0

2

4

6

8

10

bd be bf bg bh

SAR

(d

B)

Static Filter-Moving Sources

BMSS-Moving Sources

BMSS-Static Sources

StaticFilter-StaticSources

Figure 4.19: Two speaker mixtures: BSS Eval Toolbox results showing comparison

between baseline IVD implementation and BMSS system with moving speakers and

stationary speakers. See text for discussion.

Page 145: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.5. Objective Evaluation of the Proposed BMSS System 124

Time (Frames)

An

gle

(D

eg

ree

s)

500 1000 1500 2000

100

200

300

0 0.5 1

(a) Speaker tracking

500 1000 1500 20000

1

2

3

4

5

Time (Frames)

No

. o

f So

urc

es

DeterminedTrue

(b) No. of sources detected

Figure 4.20: Examples of four moving speakers - Tracking and detection performed by

the proposed BMSS system. In (a) the location expectation of speakers is depicted to

show tracking. In (a) separate detected ‘tracks’ of the four speakers in space can be

observed. Near the time frames in the middle (800 - 1200), the speakers take a long

pause and start speaking again while carrying out movements. In (b) the true speaker

activities are compared with those determined by the proposed BMSS system.

distribution representation noisy and more unpredictable over short time durations

than in a scenario with less speakers. As such, the time-frequency components are

misdirected and may come as interference in the desired speaker output leading to a

degraded performance.

The performance of convolutive source separation methods is dependent on several fac-

tors including the number of microphones, geometry of the utilised microphone array,

number of sound sources, blind or semi-blind algorithm, computational cost, rever-

beration time of the room, and positioning of the sources and the microphones. As

such, results of other techniques are included to give reader an overall perspective of

performance of the proposed BMSS system in comparison to state-of-the-art methods.

The RANSAC and expectation maximization based algorithm [24] report 2.23 dB SDR

and 7.26 dB SIR (BSS Eval measures) with four stationary speakers using three om-

nidirectional microphones (8 cm spacing). The results have been reported considering

a known number of speakers, and with simulated data including babble noise in the

background. The moving speaker tracking and separation results have been included

for two and three real world speaker mixtures, reporting 10.5 dB and 3.2 dB SIR re-

Page 146: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.6. Conclusion 125

spectively [24]. The T60 time of the target speakers and ambient speakers was 0.25s and

0.45s. Although the technique is able to perform with less computational cost, similar

to the proposed BMSS system, objective results of the proposed system pipeline de-

pict superior performance, which performs separation without knowing the number of

speakers beforehand.

A hybrid audio-visual approach to deal with moving speaker separation has been pro-

posed in [60], which performs source separation using three-dimensional movement

tracker based on Markov Chain Monte Carlo particle filter and robust least squares fre-

quency invariant data independent beam-former. A 16-component microphone array

with a radius of 0.2m has been used, along with four colour cameras to collect video

data. With a simulated reverberation time of 300ms, an average 5.49 dB SDR and 11.46

dB SIR for three stationary speaker mixtures using BSS Eval have been reported. Its

reverberation time is very similar as in the experiments performed for proposed BMSS

system; however, these results are worse than what the proposed BMSS system achieved

with four stationary sources (see static sources in Fig. 4.22). Using the 0.5s real world

recordings for moving speakers, with 130ms reverberation time, 6.63 dB average SDR

and 9.31 dB SIR have been reported. In comparison to the proposed BMSS system the

computational cost is greater as the system uses both the audio and video cues. The

applicability of such a system can also be limited for small electronic equipment or in

scenarios where video data is not readily available.

4.6 Conclusion

The baseline IVD-based source separation system was analysed for potential to perform

blind moving source separation (BMSS) in realistic acoustic scenarios. These include

dealing with time-variant mixing system parameters, in the form of number of the

speakers and their locations. In this chapter a novel IVD-based algorithm is proposed to

blindly cater for the time-variant mixing conditions. Denoising of the IVD distribution

is carried out using the von-Mises modelling in space and IIR filtering in time to

extract reliable speaker estimates over short time durations. It was found that in

order to perform speaker detection and tracking with reliability, system adaptation to

Page 147: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.6. Conclusion 126

0

1

2

3

4

5

6

7

8

9

aceg bdfh

ISR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

0

2

4

6

8

10

12

aceg bdfh

SIR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

11

11.5

12

12.5

13

13.5

14

14.5

aceg bdfh

SAR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

0

1

2

3

4

5

6

7

aceg bdfh

SDR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

29

30

31

32

33

34

35

36

37

aceg bdfh

OP

S

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

46

48

50

52

54

56

58

60

62

aceg bdfh

TPS

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

(a) (b)

(c) (d)

(e) (f)

Figure 4.21: Four speaker mixtures: PEASS Toolbox results showing comparison be-

tween baseline IVD implementation and BMSS system with moving speakers and sta-

tionary speakers. See text for discussion.

Page 148: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.6. Conclusion 127

0

2

4

6

8

10

12

14

16

18

20

aceg bdfhSI

R (

dB

)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

0

1

2

3

4

5

6

7

aceg bdfh

SAR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

0

1

2

3

4

5

6

7

aceg bdfh

SDR

(d

B)

Static Filter-Moving Sources

BMSS-MovingSources

BMSS-StaticSources

StaticFilter-StaticSources

(a) (b)

(c)

Figure 4.22: Four speaker mixtures: BSS Eval Toolbox results showing comparison

between baseline IVD implementation and BMSS system with moving speakers and

stationary speakers. See text for discussion.

Page 149: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

4.6. Conclusion 128

changing mixing conditions is critical. For example over short time durations due to

added number of speakers, that result in more reflections and increased violation of

the W-disjoint orthogonality, there is poor IVD distribution representation. Hence,

IVD distribution floor is approximated at every time frame and peak detection is done

accordingly. The proposed system adapts the threshold for IVD distribution peak

detection in order to reliably detect new speakers.

It was established that the speaker energy and its DRR dictate the speaker detection

rate and the separated output quality. A low energy, low DRR speaker has noisy IVD

distribution and poor separated output quality in the presence of comparatively more

energetic speakers. In experiments and results, speakers with relatively comparable

overall energy levels were included for objective evaluation of the system. During

natural speech activity, drastic energy variations occur over short time durations, and

movement of speakers adds to the already challenging problem. In such scenarios

the IIR filtering, adaptive thresholding, time-variant regions of interest and location

expectation algorithms distinguish false peaks from true speaker peaks.

The proposed system pipeline has been tested to perform efficiently with two and four

speaker mixtures in stationary and moving speaker scenarios. The objective evaluation

of the system has been carried out with BSS Eval and PEASS toolboxes, while inher-

ently taking into account the speaker detection and tracking capability. As opposed to

baseline IVD-based implementation, the proposed BMSS system provided with algo-

rithms to deal with practical acoustic scenarios which involve swiftly, accurately and

reliably estimating the time varying mixing parameters. The proposed BMSS system

in comparison with baseline IVD-based system performed better in moving sources sce-

nario and on par in the stationary sources scenario, with the added advantage of blindly

detecting and localising speakers. Also, the proposed BMSS system has performance

comparable to other state-of-the-art systems such as RANSAC or hybrid audio-visual

approaches in terms of SDR and SIR with the commonly used BSS Eval evaluation

toolbox.

Page 150: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Chapter 5

Conclusions and Further Work

Source separation techniques can be generally categorized into stochastic, beam-forming

and time-frequency sparsity-based algorithms. This thesis dealt with the analysis and

improvement of an IVD-based source separation algorithm developed by Gunel et al.

[11]. In comparison with other source separation techniques, the IVD-based technique

is a time-frequency sparsity-based technique which requires low processing time. It

exploits multipath characteristics of the surrounding environment to perform source

separation [11]. The technique utilises a small coincident microphone array, which is

very desirable in electronic devices such as high quality hearing aids and cell phones.

This thesis dealt with the analysis and improvement of IVD-based technique in physi-

cally stationary and moving sources scenarios.

5.1 Conclusions

During new analysis presented in Section 3.3, the IVD-based source separation system

was investigated to gain a deeper understanding and propose new algorithms wherever

shortcomings were identified. Analysis was carried out considering the microphone

array response, room response and the impact of multiple sources on the time-frequency

IVD measurements. Due to non-coincidence of the cardioid microphones comprising

the array, location-based microphone array responses revealed spatial aliasing above a

limit frequency. The microphone array responses below the limit frequency were used

as a look-up table to improve the IVD measurements. The proposed measurement-

129

Page 151: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.1. Conclusions 130

based correction technique significantly reduced the average localisation error for both

the evaluated rooms using their room impulse responses for single source scenarios (see

Section 3.4.1). It was found that the impact of correction with two or more sources was

not as significant, reason being the noisy IVD measurements due to simultaneously

active sources. The frequency resolution, while processing with short time Fourier

transform, can be improved to discriminate more among the source signals and reduce

the impact of added speakers; however, at the expense of poor time resolution it is not

feasible.

A new algorithm was proposed to utilise the microphone array geometry to improve

source separation performance. Due to tetrahedral configuration of cardioid micro-

phones comprising the array, only one or two relevant cardioid microphones were

chosen by appropriate weighting to separate the source of interest (see Section 3.5).

Such source specific cardioid weighting resulted in superior output quality as a result

of reduced non-coincidence, and suppression of unwanted reflections and overlapping

time-frequency components from other than the desired source location.

It was found that IVD mixing parameters vary from one room to another, and are

also dependent on the time-frequency content of sources. As such, a greater spread

of IVD measurements was observed with the room of greater reverberation time (see

Section 3.3.3). Also, within the same room the IVD statistics varied from one location to

another due to location-based varying mixing parameters. Based on these observations,

a new non-exhaustive spatial filter adaptation method was proposed to cater for effect

of varying target source data, its interaction with interfering sources and its interaction

with the surrounding room environment (see Section 3.6).

Experiments for objective evaluation were conducted to assess the performance of pro-

posed modifications over the baseline IVD-based source separation system, utilising

various combinations of speech and music with the widely used BSS Eval toolbox [126].

An average SDR improvement of 1.96 dB with two sources and 1.23 dB with four

sources was achieved for an ITU-R BS1116 standard listening room (T60 = 0.32 s).

An improvement of 1.91 dB for two sources and 1.21 dB for four sources in SDR was

achieved with a room of a greater reverberation time (T60 = 0.67 s). Similarly, aver-

Page 152: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.1. Conclusions 131

age SIR improvement of 2.96 dB with two sources and 3.38 dB with four sources was

achieved for an ITU-R BS1116 standard listening room (T60 = 0.32 s). An improve-

ment of 3.04 dB for two sources and 3.28 dB for four sources in SIR was achieved with

a room of greater reverberation time (T60 = 0.67 s).

In Chapter 4, dynamically changing acoustic scenarios were considered. For example,

in real life acoustic scenarios a number of speakers may enter or leave the acoustic scene,

or the speakers may move around in the surrounding environment. New analysis of the

IVD distribution in space, as presented in Section 4.3, exhibited system potential to

detect and track changes in the time varying mixing conditions. Although the raw IVD

distribution provided with inaccurate and unreliable speaker information, denoising of

the IVD distribution in time and space resulted in reliable speaker IVD distribution

representation (see Section 4.3.1).

Peaks in IVD distribution over short periods of time represented potential speaker

activity, hence peak detection algorithm was proposed to detect and track such peaks.

It was shown that as the number of speakers increase, the IVD distribution noise floor

rises and the speaker peaks become less prominent. This was observed due to added

reflections and greater overlapping time-frequency components with added number of

sources. As a result, the threshold for peak detection was found to be dependent on the

number of active speakers. Therefore, the threshold for peak detection was designed to

be adaptive (see Section 4.4.2).

Also due to the start-stop nature of speech, with shorter pauses in between words and

longer pauses in between sentences, a location expectation algorithm was designed to

deal with each speaker based on its peak detection rate (see Section 4.4.3). In the

proposed BMSS system pipeline, IVD distribution peaks were locally isolated in space

using time-variant regions of interest, and were checked for consistency by incrementing

location expectation each time a peak surpassed the threshold. As such, consistent

peaks after approximately 190 ms of activity were declared as active speakers. Similarly,

when the peaks were not detected, location expectation in corresponding region of

interest was decremented until the speaker was declared as absent.

Using the denoised IVD distribution, adaptive peak detection and location expectation

Page 153: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.2. Contributions 132

algorithms, the IVD-based system was developed into a blind moving source separation

system. The proposed BMSS system performed with effectiveness while dealing with

time-variant speaker locations and time-variant number of speakers.

5.2 Contributions

The novel contributions presented in this thesis can be summarised as follows:

1. New analysis of the IVD-based system was presented while taking into account

the microphone array response, effect of the surrounding room environment and

the impact of multiple sources on a time-frequency IVD measurement (see Section

3.3).

2. Also, with the aim of dealing with time-variant mixing parameters, new analysis

of the IVD-based system was presented considering IVD distribution in space (see

Section 4.3).

3. A new measurement-based microphone array correction algorithm was proposed

using the IVD measurements of a real microphone array, as presented in Section

3.4, and was evaluated for two different room environments while using their room

impulse responses in Section 3.4.1. As a result, the average localisation error of

a single source was reduced significantly to almost half the original error.

4. A new algorithm was proposed to select and weigh the cardioid microphones based

on desired source locations (Section 3.5). The proposed algorithm was evaluated

using the widely accepted BSS Eval toolbox, and was shown to give superior

output sound quality as compared to the baseline IVD-based system proposed by

Gunel et al. [11].

5. An algorithm was proposed to adaptively update the spatial filter parameters

based on IVD statistics of target and interfering sources (Section 3.6). The opti-

mum spatial filter directivity and beam-width was set in a non-exhaustive manner

using energy weighted mean and standard deviation of locally isolated IVD mea-

surements.

Page 154: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.3. Further Work 133

6. A new blind source separation system pipeline based on the denoised IVD dis-

tribution in space was proposed to deal with time-variant mixing conditions in

Chapter 4. The proposed system is capable of dealing with moving speakers

in a blind manner, by performing speaker detection, localisation, tracking and

eventually separation.

5.3 Further Work

The work presented in this thesis raises new ideas for further work, which are presented

as follows:

5.3.1 Objective evaluation with controlled speaker movements and

speed variations

The IVD-based source separation system has been developed to deal with practical

acoustic scenarios as part of the work presented in this thesis; however, the evaluation of

tracking results with human subjects has only been subjectively possible. Although the

speakers were directed to walk as naturally as possible in the confined space, the speed

of moving speakers could not be controlled. Also with more number of speakers, it was

difficult to carry out the repulsive step of closely located speakers. An area in further

work would be to have some controlled real world scenarios to objectively evaluate

the tracked speaker movements based on accurately known ground truth using video

cameras or precise speaker movements. Also evaluation of the separation performance

with varying speeds of speaker movements is required.

5.3.2 Dealing with co-locating and crossing over speakers

As the IVD-based source separation system relies on the direction of arrival of sound,

separation is not possible if the sources co-locate in space. As a result, after the sources

cross over, it is not possible to match the source output channels. Such issues can be

resolved by using speaker classification algorithms incorporated into the IVD-based

Page 155: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.3. Further Work 134

source separation system. The classifier can be trained on the separated speakers while

they are located away from each other, and when the speakers cross, classification of

the time-frequency components can assist in performing source separation. Similar

work has been proposed by Adigoglu et al. [148], which aims to combine source sepa-

ration and feature extraction to improve the recognition performance. Including this,

other source separation algorithms such as ICA can be combined with the IVD-based

source separation algorithm to have a hybrid technique capable of separating co-located

speakers.

Another area of further work would be to use multiple coincident microphone arrays to

have more spatial diversity. Two sources co-locating in space for one microphone array,

can be spatially separated for the other microphone array. Hence, combining data from

multiple microphone arrays can lead to completely independent source propagation

models. Also, the variation in signal power recorded by multiple microphone arrays

can add an extra dimension to the speaker tracking algorithm.

5.3.3 Accurate modelling of IVD distribution

In the literature, it is found that source statistics vary in distribution from Laplacian in

anechoic environments to von-Mises in more reverberant environments [121][11]. Due

to the use of real reverberant recordings, IVD measurements were modelled with the

von-Mises distribution function in this thesis; however, the distribution may vary over

short time durations and should ideally be modelled accordingly [121]. As such, to

achieve better output sound quality, another area of further work would be to exploit

locally isolated source IVD statistics using the proposed regions of interest (see Section

3.6), in order to model the source IVD distribution as an optimum weighting of different

distributions.

Page 156: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

5.3. Further Work 135

5.3.4 Combining the proposed stationary source separation system

with BMSS system

The proposed modifications in Chapter 3 and Chapter 4 address the stationary and

moving speaker scenarios respectively. As such they have different sets of advantages,

the former aims to improve the overall output quality while latter caters for practical

acoustic scenarios. Another area of the further work is to combine the two systems

to achieve a hybrid IVD-based source separation system. Improvement in the output

sound quality can be achieved while dealing with moving speakers by assuming sta-

tionarity over short time durations. For example by constantly updating the proposed

cardioid weight matrix (see Section 3.5) based on the tracked speaker locations. Also

the proposed moving speaker algorithms can help in speaker detection and localisation

while dealing with physically stationary speakers. For example, the proposed speaker

detection algorithm can be used to detect start-stop activity of speech for use in the

adaptive spatial filter algorithm. As a result, when the source is detected to be absent,

the adaptive spatial filter beam-width can be reduced to suppress unwanted interfer-

ence.

In summary this thesis dealt with analysis and improvement of a unique direction of

arrival based source separation algorithm which utilises signals captured by a coinci-

dent microphone array. The IVD-based system has many interesting applications in

the context of 3D signal processing. The proposed modifications have been proven to

be effective in a non-exhaustive manner, providing solutions for dealing with practi-

cal acoustic scenarios, which include moving speakers, while delivering state-of-the-art

results.

Page 157: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Appendix A

Appdx A: BSS Demonstration

Preview

This appendix presents the blind source separation application based on the proposed

algorithms. The required equipment, application set-up for the equipment and graph-

ical user interface (GUI) are presented.

A.1 Required Equipment

The equipment required to obtain audio recordings and interface with the application

are as follows:

1. Cardioid microphones in tetrahedral configuration and as coincident as possible.

2. Audio hardware responsible for analog to digital conversion and providing audio

interface.

3. A Windows, MAC or Linux personal computer is required, with software installed

to provide interface for four audio channels. Audio Sound Input Output (ASIO)

software has been utilised in the developed application to capture the four audio

signals. In current work Windows Operating System (OS) has been used for stand

alone processing, however; the application can be compiled for any platform on

which Matlab 2012b or Matlab 2014a is installed.

136

Page 158: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

A.2. GUI Layout 137

A.2 GUI Layout

The designed GUI consists of six distinct panels. These include the device panel, control

panel, rate of adaptation panel, figures panel, sources panel and the volume panel.

Figure A.1: Graphical user interface of BSS application

A.2.1 Device Connected Panel

The device connected panel allows the user to input the name of the capturing device

to be interfaced with Matlab. In order to correctly map the channels onto the software

it also provides option to set the input channel numbers of the microphone array.

A.2.2 Control Panel

Another part of the GUI is the control panel, responsible for setting the type of input

being received. It has the options of A-format or B-format microphone input. Also the

control panel lets the user select among the proposed algorithms to perform IVD based

source separation.

Page 159: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

A.2. GUI Layout 138

Figure A.2: Device connected panel

(a) Microphone Selection (b) Algorithm selection

Figure A.3: Control Panel in GUI

The algorithms include user input, localisation, adaptive spatial filters, cardioid selec-

tion, adaptive filters with cardioid selection and IVD distribution tracking. User input

is for the basic IVD source separation algorithm. It requires user input to set spatial

filter parameters including filter location and the beam-width. Localisation algorithm

is for stationary sources scenario to determine the source locations. It is equivalent to

IVD distribution tracking which on top of the localisation performs added functionality

of speaker tracking as well. Adaptive spatial filter algorithm is for adaptively updat-

ing the spatial filter parameters without requiring the user input. Cardioid selection

algorithm chooses the best cardioid microphones based on the desired source location

to have a better output quality.

Page 160: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

A.2. GUI Layout 139

A.2.3 Rate of Spatial Filter Adaptation

The rate of adaptation of spatial filter parameters can be set for subjective assessment

of the output sound quality. Though mostly used in the algorithm development, the

option is available to set the rate of adaptation of spatial filter locations and the beam-

widths. The values can be input directly or they can be varied with slide bars.

Figure A.4: Rate of spatial filter adaptation panel

A.2.4 Figures Panel

Figure A.5: Figures panel

The figures panel displays two sets of information in real time. First figure shows the

denoised IVD distribution in space with red line, and the approximated instantaneous

Page 161: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

A.2. GUI Layout 140

IVD distribution floor with a horizontal black line. The second figure depicts the

spatial filters being currently used to filter out the desired IVDs. The number of

sources currently detected by the system is shown on the top right corner of the figures

panel.

A.2.5 Sources Panel

The sources panel is used to display and select amongst the detected sources, labelled

as ‘Tracking’ in green. The undetected channels are labelled as ‘On Hold’ in yellow.

The available options can be chosen in the desired combination. For example in Fig.

A.6 the source currently at 138 degrees in being listened. The location is automatically

updated based on source tracking algorithm.

Figure A.6: Sources panel

There is a separate panel for controlling the output source power, labelled as ‘Volume’,

as can be seen in Fig. A.1.

Page 162: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Appendix B

Intensity Measurement

In terms of physics, sound is fundamentally the movement and concentration of the

air particles on a molecular level. If a point source is active in a free field, the sound

propagates in all directions and the energy density attenuates by 1/r2, where r is

the distance to source. The attenuation happens as the same amount of energy is

distributed over a larger area because the sphere grows as a function of radial distance.

A plane wave is a wave front where propagating wave is not spherical but a plane.

At a sufficient distance away from the source, spherical waves are approximated as a

plane waves [127]. As such, an acoustic wave propagating can be modelled as a simple

sinusoidal plane wave. It depicts the pressure compressions and rarefactions as the

wave progresses.

Consider a plane wave travelling with speed c in the positive x direction with respect

to a specified Cartesian axis. It can be modelled as:

y(ω, t) = A sin(kx− ωt+ φ). (B.1)

The notation with time (t) and angular frequency (ω) is dropped in the derivation to

follow for simplicity. k is the wave number, x represents distance, φ is the known initial

phase of the wave and A is the extent of displacement of the medium particles from

their equilibrium position. The velocity of mediums deformation due to compressions

and rarefactions for such travelling wave is given as:

141

Page 163: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

142

dy

dt= A

d

dtsin(kx− ωt+ φ), (B.2)

V = Aω cos(kx− ωt+ φ), (B.3)

where V represents the velocity. Let us analyse the energy density of the modelled

one dimensional sinusoidal acoustic wave travelling in an elastic medium of a constant

cross-sectional area. Consider an infinitesimal volume, ν, of the medium in which the

wave is propagating. The total energy, E, in this volume is given by:

E = Ek + Ep. (B.4)

Ek is the kinetic energy in the considered volume of the medium and Ep is the potential

energy which is due to the displacement of volume of medium from its equilibrium

position. Kinetic energy is related to mass and velocity as:

Ek =1

2mV 2. (B.5)

Ek =1

2mA2ω2 cos2(kx− ωt+ φ). (B.6)

V is replaced in equation (B.5) with equivalent given in equation (B.3) and m represents

mass. The potential energy of a rectilinear motion, as that of acoustic waves, can be

represented as the negative of the work done in displacing the infinitesimal volume from

equilibrium position over displacement y.

Ep = −∫ y

0h dy, (B.7)

Ep = −1

2h y2, (B.8)

where h is the elastic constant of the restoring force acting on the infinitesimal mass.

This force constant h can be identified by applying Newtons second law to the mass

element.

Page 164: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

143

F = m a, (B.9)

F = m Ad2

dt2sin(kx− ωt+ φ), (B.10)

F = −mω2 y, (B.11)

where a represents acceleration of the mediums deformation and F represents force.

Equation (B.11) is a form of the Hookes law [149]. It implies that the force exerted by

an elastic spring is directly opposite to the direction of displacement and is proportional

to the amount of displacement from equilibrium position. It also depends on the elastic

constant characterised by the type of spring. Objects that quickly regain their original

shape after being deformed by a force, with the molecules or atoms of their material

returning to the initial state of stable equilibrium, often obey this law. In case of

acoustic wave propagation it is assumed that after the wave has passed through a

certain point, the medium returns to the initial state of equilibrium. From equation

(B.11) elastic constant of Hookes law can be identified as:

h = −mω2. (B.12)

Using this value in equation (B.8), potential energy of the infinitesimal mass element

is:

Ep =1

2m ω2 y2. (B.13)

Ep =1

2m A2ω2 sin2(kx− ωt+ φ). (B.14)

Using equations (B.4), (B.6) and (B.14), we can write energy of the propagating wave

as:

E =1

2m A2ω2 (sin2(kx− ωt+ φ) + cos2(kx− ωt+ φ)). (B.15)

Page 165: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

144

Given sin2(kx− ωt+ φ) + cos2(kx− ωt+ φ) = 1,

E =1

2m A2ω2. (B.16)

As an ideal plane wave progresses the total energy in the infinitesimally small volume

remains constant, only the proportions of kinetic and potential energy vary. Using B.16

energy per unit volume, the energy density, is expressed as:

D =1

2ρ A2ω2, (B.17)

where D is the energy density and ρ represents density of the considered volume. We

know intensity is defined as the power delivered per unit area and power is energy per

unit time. Therefore:

I =P

Λ, (B.18)

I =E

Λ t, (B.19)

where Λ is cross-section area of the medium perpendicular to the propagation of acoustic

plane wave. I represents intensity and P is the power. Equation (B.19) can be written

in the form of volume (ν) considered and the distance, x, travelled by the sinusoidal

acoustic wave perpendicular to the area, Λ.

I =E x

ν t, (B.20)

I =D x

t. (B.21)

The rate of change of distance covered by the plane wave is its speed c and using

equation (B.17):

I =1

2ρA2ω2 c. (B.22)

Rewriting after multiplying and dividing by ρc:

Page 166: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

145

I =1

2ρc(ρcAω)2. (B.23)

Investigating the standard international (SI) units of the derived intensity in equation

(B.23):

ρc(Aω)2 =kg

m3

m2

s2→ kg

ms2. (B.24)

In sound recordings, pressure (P ) of the incoming acoustic wave is captured, given as:

P =F

Λ, (B.25)

where F represents the force exerted by the acoustic wave. Investigation of the SI units

of equation (B.25) reveals equivalence with equation (B.24). Based on the outcome

equation (B.23) can be re-written as:

P ∝ ρcAω, (B.26)

I ∝ 1

2ρcP 2. (B.27)

As derived in equation (B.27), if P 2 is measured as a vector quantity, the intensity

as well as the direction of the acoustic plane wave impinging at the microphone can

be determined. This relation of the acoustic pressure with acoustic intensity can be

represented in the time-frequency domain as:

I(ω, t) ∝ 1

2ρcP (ω, t)2. (B.28)

Page 167: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography

[1] S. Haykin and Z. Chen. The cocktail party problem. Neural Computation,

17(9):1875–1902, 2005.

[2] D. Wang and G. J. Brown. Computational Auditory Scene Analysis: Principles,

Algorithms, and Applications. Wiley-IEEE Press, 2006.

[3] L. J. Griffiths and C. W. Jim. An alternative approach to linearly con-

strained adaptive beamforming. IEEE Transactions on Antennas and Propa-

gation, 30(1):27–34, 1982.

[4] L. Parra and C. Alvino. Geometric source separation: merging convolutive source

separation with geometric beamforming. IEEE Transactions on Speech and Audio

Processing, 10(6):352–362, 2002.

[5] J. Mourjopoulos. On the variation and invertibility of room impulse response

functions. Journal of Sound and Vibration, 102(2):217–228, 1985.

[6] A. J. Van Veen. Algebraic methods for deterministic blind beamforming. Pro-

ceedings of the IEEE, 86(10):1987–2008, 1998.

[7] J. F. Cardoso. Blind signal separation: statistical principles. Proceedings of the

IEEE, 86(10):2009–2025, 1998.

[8] P. Comon. Independent component analysis, a new concept? Signal processing,

36(3):287–314, 1994.

[9] S. Rickard, R. Balan, and J. Rosca. Real-time time-frequency based blind source

separation. aje, 2:1, 2001.

146

Page 168: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 147

[10] H. Saruwatari, Y. Mori, T. Takatani, S. Ukai, K. Shikano, T. Hiekata, and

T. Morita. Two-stage blind source separation based on ICA and binary masking

for real-time robot audition system. In IEEE/RSJ International Conference on

Intelligent Robots and Systems (IROS), pages 2303–2308. IEEE, 2005.

[11] B. Gunel, H. Hacihabiboglu, and A. M. Kondoz. Acoustic source separation of

convolutive mixtures based on intensity vector statistics. IEEE Transactions on

Audio, Speech, and Language Processing, 16(4):748–756, 2008.

[12] P. Bofill and M. Zibulevsky. Underdetermined blind source separation using

sparse representations. Signal processing, 81(11):2353–2362, 2001.

[13] H. Sawada, S. Araki, and S. Makino. Underdetermined convolutive blind source

separation via frequency bin-wise clustering and permutation alignment. IEEE

Transactions on Audio, Speech, and Language Processing, 19(3):516–527, March

2011.

[14] N. Roman and J. Woodruff. Speech intelligibility in reverberation with ideal

binary masking: Effects of early reflections and signal-to-noise ratio threshold.

The Journal of the Acoustical Society of America, 133(3):1707–1717, 2013.

[15] S. Araki, H. Sawada, R. Mukai, and S. Makino. Underdetermined blind sparse

source separation for arbitrarily arranged multiple sensors. Signal Processing,

87(8):1833 – 1847, 2007. Independent Component Analysis and Blind Source

Separation.

[16] N. Madhu and J. Wouters. Localisation-based, situation-adaptive mask gener-

ation for source separation. In International Symposium on Communications,

Control and Signal Processing (ISCCSP), pages 1–6, March 2010.

[17] N. Chong, S. Wong, B. T. Vo, N. Sven, and I. Murray. Multiple moving speaker

tracking via degenerate unmixing estimation technique and cardinality balanced

multi-target multi-Bernoulli filter (DUET-CBMeMBer). In IEEE Ninth Inter-

national Conference on Intelligent Sensors, Sensor Networks and Information

Processing (ISSNIP), pages 1–6. IEEE, 2014.

Page 169: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 148

[18] S. M. Naqvi, Y. Miao, and J. A. Chambers. A multimodal approach to blind

source separation of moving sources. IEEE Journal of Selected Topics in Signal

Processing, 4(5):895–910, Oct 2010.

[19] G. Bao, Z. Ye, X. Xu, and Y. Zhou. A compressed sensing approach to blind sepa-

ration of speech mixture based on a two-layer sparsity model. IEEE Transactions

on Audio, Speech, and Language Processing, 21(5):899–906, May 2013.

[20] M. Souden, S. Araki, K. Kinoshita, T. Nakatani, and H. Sawada. A multichannel

MMSE-based framework for speech source separation and noise reduction. IEEE

Transactions on Audio, Speech, and Language Processing, 21(9):1913–1928, 2013.

[21] N. Sturmel and L. Daudet. Informed source separation using iterative reconstruc-

tion. IEEE Transactions on Audio, Speech, and Language Processing, 21(1):178–

185, Jan 2013.

[22] T. Yoshioka, T. Nakatani, M. Miyoshi, and H. G. Okuno. Blind separation and

dereverberation of speech mixtures by joint optimization. IEEE Transactions on

Audio, Speech, and Language Processing, 19(1):69–84, Jan 2011.

[23] Z. Koldovsky, J. Malek, P. Tichavsky, and F. Nesta. Semi-blind noise extraction

using partially known position of the target source. IEEE Transactions on Audio,

Speech, and Language Processing, 21(10):2029–2041, 2013.

[24] J. Traa and P. Smaragdis. Multichannel source separation and tracking with

RANSAC and directional statistics. IEEE/ACM Transactions on Audio, Speech,

and Language Processing, 22(12):2233–2243, 2014.

[25] E. C. Cherry. Some experiments on the recognition of speech, with one and with

two ears. The Journal of the Acoustical Society of America, 25(5):975–979, 1953.

[26] A. Hyvarinen and E. Oja. Independent component analysis: algorithms and

applications. Neural networks, 13(4):411–430, 2000.

[27] S. Rickard. The DUET blind source separation algorithm. In Blind Speech Sep-

aration, pages 217–241. Springer, 2007.

Page 170: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 149

[28] X. R. Cao and R. Liu. General approach to blind source separation. IEEE

Transactions on Signal Processing, 44(3):562–571, Mar 1996.

[29] L. Parra and C. Spence. Convolutive blind separation of non-stationary sources.

IEEE Transactions on Speech and Audio Processing, 8(3):320–327, May 2000.

[30] A. Cichocki and S. Amari. Adaptive blind signal and image processing: learning

algorithms and applications, volume 1. John Wiley & Sons, 2002.

[31] N. Mitianoudis and M. E. Davies. Audio source separation of convolutive mix-

tures. IEEE Transactions on Speech and Audio Processing, 11(5):489–497, 2003.

[32] P. Comon, C. Jutten, and J. Herault. Blind separation of sources, part ii: Prob-

lems statement. Signal processing, 24(1):11–20, 1991.

[33] T. W. Lee. Independent Component Analysis. Springer, 1998.

[34] V. S. James. Independent Component Analysis: A Tutorial Introduction, 2004.

[35] A. Hyvarinen, J. Karhunen, and E. Oja. Independent component analysis, vol-

ume 46. John Wiley & Sons, 2004.

[36] H. F. Trotter. An elementary proof of the central limit theorem. Archiv der

Mathematik, 10(1):226–234, 1959.

[37] H. Li and T. Adali. A class of complex ICA algorithms based on the kurtosis cost

function. IEEE Transactions on Neural Networks, 19(3):408–420, March 2008.

[38] D. MacKay. Maximum likelihood and covariant algorithms for independent com-

ponent analysis. Technical report, Citeseer, 1996.

[39] B. Perlmutter and L. Parra. Maximum likelihood blind source separation: A

context-sensitive generalization of ICA. In Neural Information Processing Sys-

tems, volume 9, pages 613–619, 1997.

[40] J. F. Cardoso. Infomax and maximum likelihood for blind source separation.

1997.

Page 171: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 150

[41] A. J. Bell and T. J. Sejnowski. An information-maximization approach to blind

separation and blind deconvolution. Neural Computation, 7(6):1129–1159, 1995.

[42] S. Amari, A. Cichocki, and H. H. Yang. A new learning algorithm for blind signal

separation. Advances in Neural Information Processing Systems, pages 757–763,

1996.

[43] T. W. Lee, M. Girolami, and T. J. Sejnowski. Independent component analysis

using an extended infomax algorithm for mixed sub-Gaussian and super-Gaussian

sources. Neural Computation, 11(2):417–441, 1999.

[44] M. Girolami and C. Fyfe. Negentropy and kurtosis as projection pursuit indices

provide generalised ICA algorithms. In Advances in Neural Information Process-

ing Systems Workshop, volume 9, 1996.

[45] A. Hyvarinen and E. Oja. A fast fixed-point algorithm for independent component

analysis. Neural Computation, 9(7):1483–1492, 1997.

[46] J. Karhunen and J. Joutsensalo. Representation and separation of signals using

nonlinear PCA type learning. Neural networks, 7(1):113–127, 1994.

[47] J. Karhunen, L. Wang, and R. Vigario. Nonlinear PCA type approaches for

source separation and independent component analysis. In IEEE International

Conference on Neural Networks, volume 2, pages 995–1000. IEEE, 1995.

[48] E. Oja. The nonlinear PCA learning rule and signal separation: Mathematical

analysis. Citeseer, 1995.

[49] S. S. Haykin. Adaptive filter theory. Pearson Education India, 2007.

[50] S. I. Amari. Natural gradient works efficiently in learning. Neural Computation,

10(2):251–276, 1998.

[51] P. Smaragdis. Blind separation of convolved mixtures in the frequency domain.

Neurocomputing, 22(1):21–34, 1998.

Page 172: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 151

[52] H. Sawada, R. Mukai, S. Araki, and S. Makino. A robust and precise method for

solving the permutation problem of frequency-domain blind source separation.

IEEE Transactions on Speech and Audio Processing, 12(5):530–538, 2004.

[53] W. Zhao, Y. Shen, P. Xu, J. Wang, Z. Yuan, Y. Wei, W. Jian, and H. Li.

A new efficient method for permutation and scaling ambiguity of blind source

separation signal blocks. In Fifth International Conference on Intelligent Control

and Information Processing (ICICIP), pages 23–28, Aug 2014.

[54] S. Ikeda and N. Murata. A method of ICA in time-frequency domain. In Pro-

ceedings ICA. Citeseer, 1999.

[55] N. Mitianoudis and M. E. Davies. Audio source separation: solutions and prob-

lems. International Journal of Adaptive Control and Signal Processing, 18(3):299–

314, 2004.

[56] L. Parra and C. D. Spence. Convolutive blind source separation using a multiple

decorrelation method, December 26 2000. US Patent 6,167,417.

[57] H. Sawada, S. Araki, R. Mukai, and S. Makino. Grouping separated frequency

components by estimating propagation model parameters in frequency-domain

blind source separation. IEEE Transactions on Audio, Speech, and Language

Processing, 15(5):1592–1604, 2007.

[58] A. Sarmiento, I. Duran-Dıaz, A. Cichocki, and S. Cruces. A contrast function

based on generalized divergences for solving the permutation problem in con-

volved speech mixtures. IEEE/ACM Transactions on Audio, Speech and Lan-

guage Processing (TASLP), 23(11):1713–1726, 2015.

[59] P. Comon and C. Jutten. Handbook of Blind Source Separation: Independent

component analysis and applications. Academic press, 2010.

[60] S. M. Naqvi, W. Wang, M.S. Khan, M. Barnard, and J. A. Chambers. Multi-

modal (audiovisual) source separation exploiting multi-speaker tracking, robust

beamforming and time-frequency masking. IET Signal Processing, 6(5):466–477,

July 2012.

Page 173: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 152

[61] J. O. Wisbeck, A. K. Barros, and R. G. Ojeda. Application of ICA in the sepa-

ration of breathing artifacts in ECG signals. 1998.

[62] A. Koutvas, E. Dermatas, and G. Kokkinakis. Blind speech separation of moving

speakers in real reverberant environments. In IEEE International Conference

on Acoustics, Speech, and Signal Processing (ICASSP), volume 2, pages II1133–

II1136 vol.2, 2000.

[63] B. D. Van Veen and K. M. Buckley. Beamforming: A versatile approach to spatial

filtering. IEEE assp magazine, 5(2):4–24, 1988.

[64] R. J. Urick. Principles of Underwater Sound for Engineers. Tata McGraw-Hill

Education, 1967.

[65] O. Yilmaz and S. Rickard. Blind separation of speech mixtures via time-frequency

masking. IEEE Transactions on Signal Processing, 52(7):1830–1847, 2004.

[66] M. Brandstein and D. Ward. Microphone arrays: signal processing techniques

and applications. Springer Science & Business Media, 2001.

[67] H. Cox, R.M. Zeskind, and M.M. Owen. Robust adaptive beamforming. IEEE

Transactions on Acoustics, Speech and Signal Processing, 35(10):1365–1376, Oct

1987.

[68] S. E. Nordholm, I. Claesson, and N. Grbic. Performance limits in subband beam-

forming. IEEE Transactions on Speech and Audio Processing, 11(3):193–203,

2003.

[69] S. Rickard and O. Yilmaz. On the approximate W-disjoint orthogonality of

speech. In IEEE International Conference on Acoustics, Speech, and Signal Pro-

cessing (ICASSP), volume 1, pages I–529–I–532, May 2002.

[70] S. Rickard and F. Dietrich. DOA estimation of many W-disjoint orthogonal

sources from two mixtures using DUET. In IEEE Proceedings of the Tenth Work-

shop on Statistical Signal and Array Processing, pages 311–314, 2000.

[71] R. O. Duda, P. E. Hart, and D. G. Stork. Pattern classification. John Wiley &

Sons, 2012.

Page 174: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 153

[72] S. Raki, S. Makino, H. Sawada, and R. Mukai. Reducing musical noise by a fine-

shift overlap-add method applied to source separation using a time-frequency

mask. In IEEE International Conference on Acoustics, Speech, and Signal Pro-

cessing (ICASSP), volume 3, pages iii–81. IEEE, 2005.

[73] J. W. Seok and K. S. Bae. Reduction of musical noise in spectral subtraction

method using subframe phase randomisation. Electronics Letters, 35(2):123–125,

1999.

[74] S. M. Naqvi, M. Yu, and J. A. Chambers. A multimodal approach to blind

source separation of moving sources. IEEE Journal of Selected Topics in Signal

Processing, 4(5):895–910, 2010.

[75] A. Hyvarinen. Fast and robust fixed-point algorithms for independent component

analysis. IEEE Transactions on Neural Networks, 10(3):626–634, 1999.

[76] R. Aichner, M. Zourub, H. Buchner, and W. Kellermann. Post-processing for con-

volutive blind source separation. In IEEE International Conference on Acoustics,

Speech and Signal Processing (ICASSP), volume 5, pages V–V. IEEE, 2006.

[77] R. Mukai, H. Sawada, S. Araki, and S. Makino. Robust real-time blind source

separation for moving speakers in a room. In IEEE International Conference

on Acoustics, Speech, and Signal Processing (ICASSP), volume 5, pages V–469.

IEEE, 2003.

[78] J. Zhang and P. C. Ching. Blind separation of moving speech sources using short-

time LOD based ICA method. In IEEE International Conference on Acoustics,

Speech and Signal Processing (ICASSP), volume 3, pages III–957. IEEE, 2007.

[79] R. E. Prieto and P. Jinachitra. Blind source separation for time-variant mixing

systems using piecewise linear approximations. In IEEE International Conference

on Acoustics, Speech, and Signal Processing (ICASSP), volume 5, pages v–301.

IEEE, 2005.

[80] W. Addison and S. Roberts. Blind source separation with non-stationary mixing

using wavelets. In Proceedings International Conference ICA, 2006.

Page 175: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 154

[81] J. Anemller and T. Gramss. On-line blind separation of moving sound sources.

ICA, 1998.

[82] K. E. Hild, D. Erdogmus, and J. C. Principe. Blind source separation of time-

varying instantaneous mixtures using an on-line algorithm. In IEEE International

Conference on Acoustics, Speech, and Signal Processing (ICASSP), volume 1,

pages I–993–I–996, May 2002.

[83] N. Chong, W. Shanhung, S. Nordholm, and I. Murray. Multiple sound source

tracking and identification via degenerate unmixing estimation technique and

cardinality balanced multi-target multi-Bernoulli filter (DUET-CBMeMBer) with

track management. In Asia-Pacific Signal and Information Processing Associa-

tion Annual Summit and Conference (APSIPA), pages 1–5, Dec 2014.

[84] J. H. DiBiase, H. F. Silverman, and M. S. Brandstein. Robust localization in

reverberant rooms. In Microphone Arrays, pages 157–180. Springer, 2001.

[85] J. P. Dmochowski, J. Benesty, and S. Affes. A generalized steered response power

method for computationally viable source localization. IEEE Transactions on

Audio, Speech, and Language Processing, 15(8):2510–2526, 2007.

[86] X. Zhao, J. Tang, L. Zhou, and Z. Wu. Accelerated steered response power

method for sound source localization via clustering search. Science China Physics,

Mechanics and Astronomy, 56(7):1329–1338, 2013.

[87] K. D. Donohue and P. M. Griffioen. Computational strategy for accelerating

robust sound source detection in dynamic scenes. In IEEE SOUTHEASTCON,

pages 1–8, March 2014.

[88] F. Rumsey. Spatial audio. CRC Press, 2012.

[89] F. L. Wightman and D. J. Kistler. The dominant role of low-frequency interaural

time differences in sound localization. The Journal of the Acoustical Society of

America, 91(3):1648–1661, 1992.

[90] K. Hamasaki, T. Nishiguchi, K. Hiyama, and K. Ono. Advanced multichannel

Page 176: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 155

audio systems with superior impression of presence and reality. In Audio Engi-

neering Society Convention 116, May 2004.

[91] K. Hamasaki, W. Hatano, K. Hiyama, S. Komiyama, and H. Okubo. 5.1 and

22.2 multichannel sound productions using an integrated surround sound panning

system. In Audio Engineering Society Convention 117, Oct 2004.

[92] P. G. Craven and M. A. Gerzon. Coincident microphone simulation covering three

dimensional space and yielding various directional outputs, August 16 1977. US

Patent 4,042,779.

[93] F. Rumsey and T. McCormick. Sound and recording: an introduction. CRC

Press, 2012.

[94] T. D. Abhayapala and D. B. Ward. Theory and design of high order sound field

microphones using spherical microphone array. In IEEE International Conference

on Acoustics, Speech, and Signal Processing (ICASSP), volume 2, pages II–1949–

II–1952, May 2002.

[95] C. Faller and M. Kolundzija. Design and limitations of non-coincidence correction

filters for Soundfield microphones. In Audio Engineering Society Convention 126.

Audio Engineering Society, 2009.

[96] A. Politis and V. Pulkki. Broadband analysis and synthesis for directional audio

coding using a-format input signals. In Audio Engineering Society Convention

131, Oct 2011.

[97] E. Benjamin and T. Chen. The native B-format microphone. In Audio Engineer-

ing Society Convention 119, Oct 2005.

[98] J. Pernaux, P. Boussard, and J. Jot. Virtual sound source positioning and mixing

in 5.1 implementation on the real-time system genesis. In Conference Proceedings

Digital Audio Effects (DAFx-98), pages 76–80. Citeseer, 1998.

[99] V. Pulkki, A. Politis, G. D. Galdo, and A. Kuntz. Parametric spatial audio

reproduction with higher-order B-format microphone input. In Audio Engineering

Society Convention 134, May 2013.

Page 177: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 156

[100] Wikispaces. First order B-format signals, 2015.

[101] J. Merimaa and V. Pulkki. Spatial impulse response rendering i: Analysis and

synthesis. Journal of the Audio Engineering Society, 53(12):1115–1127, 2005.

[102] A. Hyvarinen. Survey on independent component analysis. Neural Computing

Surveys, 2(4):94–128, 1999.

[103] M. R. Portnoff. Time-frequency representation of digital signals and systems

based on short-time Fourier analysis. IEEE Transactions on Acoustics, Speech

and Signal Processing, 28(1):55–69, Feb 1980.

[104] Z. Koldovsky, P. Tichavsky, and E. Oja. Efficient variant of algorithm fastICA for

independent component analysis attaining the Cramer-Rao lower bound. IEEE

Transactions on Neural Networks, 17(5):1265–1277, 2006.

[105] S. C. Douglas, M. Gupta, H. Sawada, and S. Makino. Spatio–temporal fastICA

algorithms for the blind separation of convolutive mixtures. IEEE Transactions

on Audio, Speech, and Language Processing, 15(5):1511–1520, 2007.

[106] H. Sawada, S. Araki, R. Mukai, and S. Makino. Blind extraction of dominant

target sources using ICA and time-frequency masking. IEEE Transactions on

Audio, Speech, and Language Processing, 14(6):2165–2173, Nov 2006.

[107] W. Wang, S. Sanei, J. A. Chambers, et al. Penalty function-based joint diagonal-

ization approach for convolutive blind separation of nonstationary sources. IEEE

Transactions on Signal Processing, 53(5):1654–1669, 2005.

[108] W. Wang, J. A. Chambers, and S. Sanei. A novel hybrid approach to the per-

mutation problem of frequency domain blind source separation. In Independent

Component Analysis and Blind Signal Separation, pages 532–539. Springer, 2004.

[109] Q. Liu, W. Wang, and P. Jackson. Use of bimodal coherence to resolve the

permutation problem in convolutive BSS. Signal Processing, 92(8):1916–1927,

2012.

Page 178: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 157

[110] J. Traa and P. Smaragdis. Blind multi-channel source separation by circular-linear

statistical modeling of phase differences. In IEEE International Conference on

Acoustics, Speech and Signal Processing (ICASSP), pages 4320–4324, May 2013.

[111] M. Mandel, R. J. Weiss, D. P. W. Ellis, et al. Model-based expectation-

maximization source separation and localization. IEEE Transactions on Audio,

Speech, and Language Processing, 18(2):382–394, 2010.

[112] A. Alinaghi, P. Jackson, Q. Liu, and W. Wang. Joint mixing vector and binau-

ral model based stereo source separation. IEEE/ACM Transactions on Audio,

Speech, and Language Processing, 22(9):1434–1448, 2014.

[113] H. E. Bree, W. F. Druyvesteyn, E. Berenschot, and M. Elwenspoek. Three-

dimensional sound intensity measurements using microflown particle velocity sen-

sors. 1999.

[114] B. Gunel, H. Hacıhabiboglu, and A. M. Kondoz. Wavelet packet based analysis

of sound fields in rooms using coincident microphone arrays. Applied Acoustics,

68(7):778–796, 2007.

[115] B. G. Hacihabiboglu, H. Hacihabiboglu, and A. Kondoz. Acoustic source sepa-

ration, October 17 2008. US Patent App. 12/734,195.

[116] R. Gatto and S. R. Jammalamadaka. The generalized von-Mises distribution.

Statistical Methodology, 4(3):341–353, 2007.

[117] Chen, X. and Wang, W. and Wang, Y. and Zhong, X. and Alinaghi, A. Rever-

berant speech separation with probabilistic time-frequency masking for B-format

recordings. Speech Communication, 68:41–54, 2015.

[118] G. Cengarle and T. Mateos. Comparison of anemometric probe and tetrahedral

microphones for sound intensity measurements. In Audio Engineering Society

Convention 130, May 2011.

[119] K. Farrar. Soundfield microphone. Wireless World, 85(1526):48–50, 1979.

Page 179: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 158

[120] M. Shujau, C.H. Ritz, and I.S. Burnett. Separation of speech sources using an

acoustic vector sensor. In International Workshop on Multimedia Signal Process-

ing (MMSP), pages 1–6, Oct 2011.

[121] X. Chen, A. Alinaghi, X. Zhong, and W. Wang. Acoustic vector sensor based

speech source separation with mixed Gaussian-Laplacian distributions. In 18th

International Conference on Digital Signal Processing (DSP), pages 1–5. IEEE,

2013.

[122] L. Moskowitz. Introduction: Tetramic single point stereo & surround sound

microphone. Online at http://www.core-sound.com/TetraMic/1.php.

[123] S. W. Smith. Digital Signal Processing: A Practical Guide for Engineers and

Scientists. Newnes, 2003.

[124] A. Farina. Advancements in impulse response measurements by sine sweeps. In

Audio Engineering Society Convention 122. Audio Engineering Society, 2007.

[125] CD Bang&Olufsen. 101. Music for Archimedes, 1992.

[126] E. Vincent, R. Gribonval, and C. Fevotte. Performance measurement in blind

audio source separation. IEEE Transactions on Audio, Speech and Language

Processing, 14(4):1462–1469, 2006.

[127] L. E. Kinsler, A. R. Frey, A. B. Coppens, and J. V. Sanders. Fundamentals of

acoustics. Wiley-VCH, 1999.

[128] J. Y. Chung. Cross-spectral method of measuring acoustic intensity without error

caused by instrument phase mismatch. The Journal of the Acoustical Society of

America, 64(6):1613–1616, 1978.

[129] J. Merimaa. Applications of a 3-D microphone array. In Audio Engineering

Society Convention 112, Apr 2002.

[130] V. Hansen and G. Munch. Making recordings for simulation tests in the

archimedes project. Journal of the Audio Engineering Society, 39(10):768–774,

1991.

Page 180: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 159

[131] N. Mitianoudis and T. Stathaki. Batch and online underdetermined source sep-

aration using Laplacian mixture models. IEEE Transactions on Audio, Speech

and Language Processing, 15(6):1818–1832, 2007.

[132] P. Bofill and M. Zibulevsky. Blind separation of more sources than mixtures using

sparsity of their short-time Fourier transform. In Proceedings ICA, volume 2000,

pages 87–92, 2000.

[133] A. Farina. A-format to b-format conversion. Online at http://pcfarina. eng.

unipr. it/Public/B-Format/A2B-conversion/A2B. htm, 2006.

[134] J. H. Zar. Biostatistical Analysis. Pearson Education India, 1999.

[135] S. W. Smith. The Scientist and Engineers Guide to Digital Signal Processing.

1997.

[136] S. Yamamoto, K. Nakadai, M. Nakano, H. Tsujino, J. M. Valin, K. Komatani,

T. Ogata, and H. G. Okuno. Real-time robot audition system that recognizes

simultaneous speech in the real world. In IEEE/RSJ International Conference

on Intelligent Robots and Systems, pages 5333–5338, Oct 2006.

[137] A. Plinge, M. H. Hennecke, and G. A. Fink. Reverberation-robust online multi-

speaker tracking by using a microphone array and CASA processing. In Proceed-

ings of International Workshop on Acoustic Signal Enhancement (IWAENC),

pages 1–4, Sept 2012.

[138] V. Pulkki. Spatial sound reproduction with directional audio coding. Journal of

the Audio Engineering Society, 55(6):503–516, 2007.

[139] K. Nakadai, G. Ince, K. Nakamura, and H. Nakajima. Robot audition for dy-

namic environments. In IEEE International Conference on Signal Processing,

Communication and Computing (ICSPCC), pages 125–130, Aug 2012.

[140] W. K. Ma, T. H. Hsieh, and C. Y. Chi. DOA estimation of quasi-stationary signals

with less sensors than sources and unknown spatial noise covariance: A Khatri-

Rao subspace approach. IEEE Transactions on Signal Processing, 58(4):2168–

2180, April 2010.

Page 181: Adaptive Blind Source Separation Based on Intensity Vector ...epubs.surrey.ac.uk/810208/1/thesis.pdf · response and/or any target source signal information. A system that approximates

Bibliography 160

[141] E. Habets. Multi-channel speech dereverberation based on a statistical model of

late reverberation. In Proceedings IEEE International Conference on Acoustics,

Speech, and Signal Processing (ICASSP), volume 4, pages iv/173–iv/176 Vol. 4,

March 2005.

[142] V. Tyagi, H. Bourlard, and C. Wellekens. On variable-scale piecewise stationary

spectral analysis of speech signals for ASR. Speech Communication, 48(9):1182 –

1191, 2006.

[143] N. Mitianoudis. A generalized directional laplacian distribution: Estimation, mix-

ture models and audio source separation. IEEE Transactions on Audio, Speech,

and Language Processing, 20(9):2397–2408, 2012.

[144] K. V. Mardia and P. E. Jupp. Directional statistics, volume 494. John Wiley &

Sons, 2009.

[145] H. Kameoka, T. Nakatani, and T. Yoshioka. Robust speech dereverberation

based on non-negativity and sparse nature of speech spectrograms. In IEEE

International Conference on Acoustics, Speech and Signal Processing (ICASSP),

pages 45–48, April 2009.

[146] N. Yoder. Peakfinder. Online at http://www.mathworks.com/matlabcentral

/fileexchange/25500-peakfinder-x0–sel–thresh–extrema–includeendpoints–

interpolate.

[147] V. Emiya, E. Vincent, N. Harlander, and V. Hohmann. Subjective and objective

quality assessment of audio source separation. IEEE Transactions on Audio,

Speech, and Language Processing, 19(7):2046–2057, 2011.

[148] K. Adiloglu and E. Vincent. A general variational bayesian framework for robust

feature extraction in multisource recordings. In IEEE International Conference

on Acoustics, Speech and Signal Processing (ICASSP), pages 273–276, March

2012.

[149] J. O. Smith. Physical Audio Signal Processing: For Virtual Musical Instruments

and Audio Effects. W3K Publishing, 2010.