rand feind

29
Thunderstorm In Situ Measurements from the Armored T28 Aircraft: Classification of 2D Probe Hydrometeor Images Rand Feind

Upload: nguyet

Post on 12-Jan-2016

40 views

Category:

Documents


0 download

DESCRIPTION

Thunderstorm In Situ Measurements from the Armored T28 Aircraft: Classification of 2D Probe Hydrometeor Images. Rand Feind. Overview. The problem The data Classes Feature Selection Classifiers Results. The Problem. Sensor on T-28 collects images - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Rand Feind

Thunderstorm In Situ Measurements from the Armored T28 Aircraft:

Classification of 2D Probe Hydrometeor Images

Rand Feind

Page 2: Rand Feind

Overview

• The problem

• The data

• Classes

• Feature Selection

• Classifiers

• Results

Page 3: Rand Feind

The Problem

• Sensor on T-28 collects images

• Number of images can number in the hundreds of thousands

• Hydrometeor classification provides key to estimating cloud characteristics

Page 4: Rand Feind

2DC and HVPS Probes

Page 5: Rand Feind

The Data

• Strips taken from the sensor

• Treated as black/white images

• 2000 images extracted for training and testing

Page 6: Rand Feind

The Classes

1. Drops - smooth perimeters; appear to be circular

2. Snow - irregular, convoluted perimeters

3. Hail - somewhat rough, lumpy perimeters; appear to be circular

4. Columns - linear like needles but wider; can have rough perimeters

5. Needles - linear and narrow

6. Dendrites - like snow but evidence of 6-way symmetry

7. Plates - appear to be planar and 6-sided

8. Holes - anomalous images (due to probe tip shedding)

Page 7: Rand Feind
Page 8: Rand Feind
Page 9: Rand Feind

Feature Selection

• Need to select features for classification

• How many?

– Literature search for ideas

– Start with many, eliminate (started with 25)

• Elimination using divergence measure– Provided base set of 6

• Trial and error– Add one at a time, check error

– Delete one of 6, check error

Page 10: Rand Feind

Example : Basic Metrics

• X Dimension - The width of the image in pixels along the flight direction (x-dimension) (e.g., the horizontal dimension).

• Y Dimension - The height of the image in pixels perpendicular to the flight direction (y-dimension) (e.g., the vertical dimension). Note: In the case of the T-28, this orientation is perpendicular to the wingspan.

• Heymsfield Diameter - The larger of the X Dimension and the Y Dimension above.

Page 11: Rand Feind

Basic Metrics

• Bottom Occulted - If the 32 photodetectors in the 2DC probe are numbered 1 through 32, this metric is the number of times photodetector 1 is shaded (i.e., the number of image pixels along the bottom edge of the image window).

• Top Occulted - Same as Bottom Occulted except photodetector 32 or the top edge of the image window.

• Total Occulted – Sum of the previous two features. Used as a particle rejection criterion in Holroyd, 1987.

Page 12: Rand Feind

Basic Metrics

• Pixel Area – Sum of the number of pixels comprising a 2D image.

• Area – Area of the particle image in square micrometers (um).

• Streak – Ratio of the x-dimension to the y-dimension. Used to detect anomalous images resulting from the shedding of droplets, from the probes tips, that are moving slower than the air stream.

Page 13: Rand Feind

Basic Metrics

• Perimeter – The perimeter is determined in three different ways each of which has a unique value. The first is determined by subtracting from the original particle image an eroded version of it . The second is determined by subtracting the original particle image from a dilated version of it. The second perimeter is always larger than the first. A perimeter or bug finding algorithm (Ballard and Brown, 1982) determines the third perimeter. The bug finding algorithm also provides an ordered sequence of coordinates around the perimeter which is used in the calculation of Fourier Descriptors.

• Maximum Area – Area of a circle using maximum length as the diameter.

Page 14: Rand Feind

Divergence

• Jeffries-Matusita (JM) distance – Values range 0 (identical) to 2 (little overlap)– Hope : one feature gives value of about 2 for

each pair of classes. Never happens.– Assumes normal distribution

dxxpxpJx

jiij

2||

Page 15: Rand Feind

Divergence cont.

• where p(x|wi) and p(x|wj) are the normal probability distributions for the two classes i and j

Page 16: Rand Feind

Features selected

• PDA– Perimeter Diameter Area (PDA) – The product of the perimeter

and diameter divided by the area. Smooth, circular images give smaller values while irregular ones give larger values.

• Linearity– Linearity – The correlation coefficient for the regression.

Values for linear images, such as of a needle or column, are closer to 1. This is as opposed to circularly symmetric images that have values closer to 0. A Holroyd measure.

• equivalent circle– Equivalent Circle – The diameter of a circle that has the same area

as the particle image.

Page 17: Rand Feind

Features selected

• Concavity– The ratio of the number of concave perimeter points to

the distance around the convex hull. Convex images give zero or small values while images with concavities give larger values.

• projection fit- The standard error of a least squares quadratic regression

of the projection of the number of pixels in the vertical along the horizontal. Smooth, circular images give low standard errors while irregular shapes give high errors.

Page 18: Rand Feind

Features selected

• convex hull– The convex hull is the distance around the

perimeter of a particle image as though a rubber band were stretched around it, or the distance traversed by rolling the image along a straight line.

Page 19: Rand Feind

Feature Distributions

• Distributions are not always:– Gaussian– Monomodal– Well separated between/among classes

Page 20: Rand Feind

Classification Methodologies

•Mahalanobis Minimum Distance

•Fuzzy Logic

•Backpropagation Neural Network

Page 21: Rand Feind

Mahalanobis

• Form of Maximum Likelihood Classifier– Assume equal a priori probabilities– A Euclidean distance with directionality

Richards, 1986

Page 22: Rand Feind

Mahalanobis(2D feature space)

Richards, 1986

Feature 1

Feature 2

P

Probability of image belonging to each class

Class 1

Class 2

Class 3

Page 23: Rand Feind

Results

• Performance or accuracy of each of 3 classifiers was derived using a separate set of training and testing sample images

Page 24: Rand Feind

Confusion Matrices

Page 25: Rand Feind

Table 1. Mahalanobis Classifier Confusion Matrix

a) Normalized by Rows - P(class j | classification i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk # Sam Drops 97.1 0.0 2.2 0.0 0.0 0.0 0.0 0.0 0.7 137 Snow 0.0 75.1 4.7 2.3 0.0 16.9 0.0 0.0 0.9 213 Hail 33.1 1.7 55.1 0.8 0.0 0.8 6.8 1.7 0.0 118 Cols 0.0 3.7 0.0 85.2 1.9 3.7 4.6 0.0 0.9 108 Need 0.0 0.0 0.0 1.7 93.1 0.0 0.0 1.7 3.4 58 Dend 0.0 30.3 6.1 6.1 0.0 57.6 0.0 0.0 0.0 33 Plates 25.9 0.0 9.8 0.7 0.0 0.7 58.0 4.9 0.0 143 Holes 0.0 0.0 0.0 0.0 0.0 0.0 0.7 96.5 2.8 144

b) Normalized by Columns - P(classification j | class i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk Drops 63.6 0.0 3.2 0.0 0.0 0.0 0.0 0.0 10.0 Snow 0.0 90.9 10.6 4.9 0.0 59.0 0.0 0.0 20.0 Hail 18.7 1.1 69.1 1.0 0.0 1.6 8.2 1.3 0.0 Cols 0.0 2.3 0.0 90.2 3.6 6.6 5.2 0.0 10.0 Need 0.0 0.0 0.0 1.0 96.4 0.0 0.0 0.7 20.0 Dend 0.0 5.7 2.1 2.0 0.0 31.1 0.0 0.0 0.0 Plates 17.7 0.0 14.9 1.0 0.0 1.6 85.6 4.7 0.0 Holes 0.0 0.0 0.0 0.0 0.0 0.0 1.0 93.3 40.0 # Sam 209 176 94 102 56 61 97 149 10

Total Acc: 78.1

Features – PDA, LINEARITY, CONVEX, EQUIV_CIR, CONCAVE, PROJECTION,STREAK

Page 26: Rand Feind

Table 2. Fuzzy Logic Classifier Confusion Matrix

a) Normalized by Rows - P(class j | classification i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk # Sam Drops 65.7 0.0 11.7 7.3 0.0 0.0 11.7 3.6 0.0 137 Snow 0.0 62.0 0.9 5.6 0.0 31.5 0.0 0.0 0.0 213 Hail 11.0 2.5 39.8 14.4 0.0 5.9 21.2 5.1 0.0 118 Cols 0.0 1.9 1.9 80.6 5.6 6.5 3.7 0.0 0.0 108 Need 0.0 0.0 0.0 1.7 94.8 0.0 0.0 3.4 0.0 58 Dend 0.0 30.3 3.0 15.2 0.0 51.5 0.0 0.0 0.0 33 Plates 11.2 0.0 23.8 26.6 0.0 0.7 32.9 4.9 0.0 143 Holes 0.0 0.0 0.7 5.6 0.0 6.3 1.4 86.1 0.0 144

b) Normalized by Columns - P(classification j | class i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk Drops 75.6 0.0 15.5 5.6 0.0 0.0 17.0 3.5 0.0 Snow 0.0 89.8 1.9 6.7 0.0 62.0 0.0 0.0 0.0 Hail 10.9 2.0 45.6 9.6 0.0 6.5 26.6 4.2 0.0 Cols 0.0 1.4 1.9 48.9 9.8 6.5 4.3 0.0 0.0 Need 0.0 0.0 0.0 0.6 90.2 0.0 0.0 1.4 0.0 Dend 0.0 6.8 1.0 2.8 0.0 15.7 0.0 0.0 0.0 Plates 13.4 0.0 33.0 21.3 0.0 0.9 50.0 4.9 0.0 Holes 0.0 0.0 1.0 4.5 0.0 8.3 2.1 86.1 0.0 # Sam 119 147 103 178 61 108 94 144 0

Total Acc: 62.8

Features – PDA, LINEARITY, CONVEX, EQUIV_CIR, CONCAVE, PROJECTION,STREAK

Page 27: Rand Feind

Table 3. Backpropagation Neural Network Classifier Confusion Matrix

a) Normalized by Rows - P(class j | classification i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk # Sam Drops 86.9 0.0 12.4 0.0 0.0 0.0 0.7 0.0 0.0 137 Snow 0.0 93.0 2.3 2.3 0.0 2.3 0.0 0.0 0.0 213 Hail 18.6 1.7 72.0 2.5 0.0 0.0 5.1 0.0 0.0 118 Cols 0.0 0.0 0.0 97.2 1.9 0.0 0.0 0.9 0.0 108 Need 0.0 0.0 0.0 3.4 93.1 0.0 0.0 3.4 0.0 58 Dend 0.0 60.6 6.1 6.1 0.0 24.2 0.0 3.0 0.0 33 Plates 13.3 0.0 9.8 0.7 0.0 0.7 69.9 5.6 0.0 143 Holes 0.0 0.0 0.0 0.7 0.0 0.0 0.0 99.3 0.0 144

b) Normalized by Columns - P(classification j | class i) x 100

Classification i:Class j Drops Snow Hail Cols Need Dend Plates Holes Unk Drops 74.4 0.0 13.8 0.0 0.0 0.0 0.9 0.0 0.0 Snow 0.0 90.0 4.1 4.2 0.0 35.7 0.0 0.0 0.0 Hail 13.8 0.9 69.1 2.5 0.0 0.0 5.6 0.0 0.0 Cols 0.0 0.0 0.0 88.2 3.6 0.0 0.0 0.6 0.0 Need 0.0 0.0 0.0 1.7 96.4 0.0 0.0 1.3 0.0 Dend 0.0 9.1 1.6 1.7 0.0 57.1 0.0 0.6 0.0 Plates 11.9 0.0 11.4 0.8 0.0 7.1 93.5 5.2 0.0 Holes 0.0 0.0 0.0 0.8 0.0 0.0 0.0 92.3 0.0 # Sam 160 220 123 119 56 14 107 155 0

Total Acc: 85.1

Features – PDA, LINEARITY, CONVEX, EQUIV_CIR, CONCAVE, PROJECTION,STREAK

Page 28: Rand Feind

What is the best classification methodology?

Page 29: Rand Feind

Conclusions

• For these samples, BPNN provides best performance; however, MMDC is a close second

• Feature set selection is tantamount to classifier selection