1 graph embedding (ge) & marginal fisher analysis (mfa) 吳沛勳...
TRANSCRIPT
![Page 1: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/1.jpg)
1
Graph Embedding (GE) &Marginal Fisher Analysis (MFA)
![Page 2: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/2.jpg)
2
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis
5. Experiments Result
![Page 3: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/3.jpg)
3
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis
5. Experiments Result
![Page 4: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/4.jpg)
4
1. Introduction
We present a general framework called Graph
Embedding (GE).
In graph embedding, the underlying merits and
shortcomings of different dimensionality reduction
schemes, existing or new, are revealed by
differences in the design of their intrinsic and penalty
graphs and their types of embedding.
A novel dimensionality reduction algorithm, Marginal
Fisher Analysis (MFA).
![Page 5: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/5.jpg)
5
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis
5. Experiments Result
![Page 6: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/6.jpg)
6
2. Face Recognition Flowchart
N : # Image (200, 20 image pre person)
: # People (10)
m : Image Size (24x24)
cN
mm
200)2424(
1][ RRxxX Nm
N
37)2424( RRw mm
w : Unitary Linear Projection Vector
k : k nearest neighbors
X
1. Training Image Set: X
2.1 MFA Space Creation: w
2.2 Projection to MFA Space: Y
20037' , RRYXwY NmT
3. k-NN Classification
1. Test image: xtest
Classification Result
Training Process
Test Process
2. Projection to MFA Space: ytest
21, kk
1)2424( Rxtest
137, Ryxwytesttest
T
test
k
winfluence
w
Y
k1 : k-NN for intrinsic graphk2 : k-NN for penalty graph
![Page 7: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/7.jpg)
7
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis
5. Experiments Result
![Page 8: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/8.jpg)
8
3. Graph Embedding
For a dimensionality reduction problem, we require an intrinsic graph G and, optionally, a penalty graph as input.
We now introduce the dimensionality reduction problem from the new point of view of graph embedding.
Let G={X,W} be an undirected weighted graph (two-way direction) with vertex set X (N nodes) and similarity (or weighted) matrix .
W :
1. Symmetric matrix
2. May be negative
44
4321
RW
xxxxX
NxxxX 21
NNRW
![Page 9: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/9.jpg)
9
3. Graph Embedding: Laplacian Matrix
L=Degree-Adjacent=>
jiijii iWD
WDL
.,(2)
0011
0001
1001
1110
W
2000
0100
0020
0003
D
2011
0101
1021
1113
L
W is weighted matrix ,also call similarity matrix
G={X,W} :
![Page 10: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/10.jpg)
10
3.1 Cost Function (1/2) Our graph-preserving criterion is given as follows:
For larger (positive) similarity samples and :
For smaller (negative) similarity samples and :
B typically is the Laplacian matrix of a penalty graph .
Use Lagrange multipliers to solve:
LyyWyyy T
dByyijji
dByy TT minarg minarg*
2(3)
2
ji yy
2
ji yy
ixjx
ix jx
pG
ppp WDLB
YLYBy
u
dBYYLYYu TT
1
rmPenalty te
graphpenalty
tion termRegulariza
graph intrinsic
0
)(
Y must be an eigenvector of the LB 1
Intrinsic Graph
Penalty Graphunknown Known
![Page 11: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/11.jpg)
11
3.1 Cost Function: Intrinsic Graph in LDA
2
100000
02
10000
004
3000
0004
300
00004
30
000004
3
66RD
ji
jicc
c
cc
ij
cc
cc
nW
ji
ji
0
1,
,
21
21
654321
]65[ ]4321[
2 4 2 6
cc
nnNN
xxxxxxX
c
If we have 6 images for 2
people.
2
1
2
10000
2
1
2
10000
004
3
4
1
4
1
4
1
004
1
4
3
4
1
4
1
004
1
4
1
4
3
4
1
004
1
4
1
4
1
4
3
66RL
02
10000
2
100000
0004
1
4
1
4
1
004
10
4
1
4
1
004
1
4
10
4
1
004
1
4
1
4
10
66RW
WDL is not important, because in GE i always not equal to j.iiW
![Page 12: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/12.jpg)
12
3.1 Cost Function: Penalty Graph
We define an intrinsic graph to be the graph G itself
Penalty graph :
1. As a graph whose vertices X are the same as those
of G.
2. Whose edge weight matrix corresponds to the similarity
characteristic that is to be suppressed in the dimension-
reduced feature space .
3. Penalty graph = constraint
},{ pp WXG
pW
![Page 13: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/13.jpg)
13
3.1 Cost Function: Penalty Graph in LDA
100000
010000
001000
000100
000010
000001
66RD p
21
21
654321
]65[ ]4321[
2 4 2 6
cc
nnNN
xxxxxxX
c
If we have 6 images for 2
people.
6
5
6
1
6
1
6
1
6
1
6
16
1
6
5
6
1
6
1
6
1
6
16
1
6
1
6
5
6
1
6
1
6
16
1
6
1
6
1
6
5
6
1
6
16
1
6
1
6
1
6
1
6
5
6
16
1
6
1
6
1
6
1
6
1
6
5
B 66R
6
1
6
1
6
1
6
1
6
1
6
16
1
6
1
6
1
6
1
6
1
6
16
1
6
1
6
1
6
1
6
1
6
16
1
6
1
6
1
6
1
6
1
6
16
1
6
1
6
1
6
1
6
1
6
16
1
6
1
6
1
6
1
6
1
6
1
66RW p
NW p
ij1
ppp WIWB DMaximize class covariance is equal to maximize data covariance.
Here, LDA penalty graph is as PCA intrinsic graph: Consider only btw-class
scatter.
![Page 14: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/14.jpg)
14
3.2 Linearization
Assuming that the low-dimensional vector representations of the
vertices can be obtained from a linear projection as ,
where w is the unitary projection vector, the objective function in
(3) becomes :
wXLXwWxwxww TT
dwwdwXBXwji
ijjT
iT
dwwdwXBXw
T
TT
T
TT
or
2
or
* minargminarg (4)
Solution
:wLwBBwLw
w
u
dwXBXwwXLXwu TTTT
10
)(
w can be solved by singular value decomposition (SVD)
![Page 15: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/15.jpg)
15
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis (MFA)
5. Experiments Result
![Page 16: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/16.jpg)
16
4.1 Marginal Fisher Analysis (MFA) (1/3)
Fig. 4.
The adjacency relationships of the intrinsic and penalty graphs for the Marginal Fisher Analysis algorithm.
k-NN : k1=5
In same class
k-NN : k2=2
In difference class
Marginal k-NN
(Within-Class) (Btw-Class)
![Page 17: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/17.jpg)
17
(15)wXWDXw
wXWDXw
S
Sw
TppT
TT
wp
c
w )(
)(minarg~
~minarg*
Cost function:
Minimize within-class (intrinsic graph)Maximize between- class (penalty graph)
(13)
else ,0
)(or )( if,1
2
~
11
11
2
iNjjNiW
wW)XX(Dw
WxwxwS
kkij
TT
iji (i)N(j) or jNi
jT
iT
c
kk
Intrinsic graph
4.1 Marginal Fisher Analysis (MFA) (2/3)
indicates the index set of the k1 nearest neighbors of the sample xi in the same
class.
)(1
iNk
)(1
iNk
L
k-NN : k1=5
In same class
![Page 18: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/18.jpg)
18
4.1 Marginal Fisher Analysis (MFA) (3/3)
(14)
else ,0
)(),(or )(),( if,1
2
~
22
22
2
jkikpij
TppT
iji (cj)P) or (i,j)(cP(i,j)
jT
iT
p
cPjicPjiW
w)XWX(Dw
WxwxwSkik
Penalty graph
How to decide k1 and k2-nearest neighbor : k-nearest neighbor (k-NN)
is a set of data pairs that are the k2 nearest pairs among the set
denote the index set belonging to the cth class
)(2
cPk}j,ij),{(i, c c
c
B
![Page 19: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/19.jpg)
19
4.1 Q&A (1/2)
Q1: MFA: How to Decide k1 k2
A1: k1 : (Within Class) Sampled five values between two and {minc(nc-1)},
and chose the value with the best MFA performance.
nc: # of images per class (subject)
(We direct use 5)
k2 : (Btw-Class) Choose the best k2 between 20 and 8Nc at sampled
intervals of 20.
Nc: # of classes (subjects)
(We direct use 20)
![Page 20: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/20.jpg)
20
MFA: Comparison with LDA Advantages :
1. The number of available projection directions (axes) is much larger than that of LDA (MFA finds more significant axes and has better classification results. ).
MFA: Rank (B-1L) LDA: Nc-1
2. There is no assumption on the data distribution, thus it is more general for discriminant analysis, LDA assumption data is approximately Gaussian distributed.
Data distribution: MFA: Non-linear LDA: Linear
3. The inter-class margin can better characterize the separability of different classes than the inter-class scatter in LDA.
MFA: Maximize margin LDA: Difference between means
Disadvantage:
LDA-> incremental LDA, MFA->?
4.1 Q&A (2/2)
margin
Positive Negative
![Page 21: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/21.jpg)
2121
Outline
1. Introduction
2. System Flowchart
3. Dimensionality Reduction- Graph Embedding 3.1 Cost Function
-Intrinsic Graph/Penalty Graph
3.2 Linearization
3.3 Example: LDA
4. Marginal Fisher Analysis
5. Experiments Result
![Page 22: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/22.jpg)
22
5. Experiments: Database (1/2)
Dadabase Yale B
People 10
#Image30 per person, random select 20 image for training, and
remain 10 image for test (G30/P20)
Image size 24x24
Variations Variable illumination, cropped face
24x24
![Page 23: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/23.jpg)
23
5. Experiments Result
k-NN k=1 k=3 k=5
ErrorRate 11.72%±3.58 11.04%±3.66 11.10±3.27
For each k run 100 times, and calculate mean and standard
deviation .
![Page 24: 1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 seanpsw@gmail.comseanpsw@gmail.com 劉冠成 zeter168@gmail.comzeter168@gmail.com 韓仁智 jchan@visionatics.com.twjchan@visionatics.com.tw](https://reader034.vdocuments.mx/reader034/viewer/2022050708/56649e585503460f94b50c70/html5/thumbnails/24.jpg)
24
Reference
1. P. Belhumeur, J. Hespanha, and D. Kriegman, “Eigenfaces vs. Fisherfac
es: Recognition Using Class Specific Linear Projection,” IEEE Trans. on
Pattern Analysis and Machine Intelligence, Vol. 19, No. 7, pp. 711–720, 1
997.
2. T.K. Kim, S.F. Wong, B. Stenger, J. Kittler and R. Cipolla, “Incremental
Linear Discriminant Analysis Using Sufficient Spanning Set
Approximations”, CVPR, pp. 1-8, 2007.
3. S. Yan, D. Xu, B. Zhang, H. Zhang, Q. Yang, and S. Lin, “Graph
Embedding and Extensions: A General Framework for Dimensionality
Reduction,” IEEE Trans. on Pattern Analysis and Machine Intelligence,
Vol. 29, No. 1, pp. 40–51, 2007.