meshnet: mesh neural network for 3d shape representation · 2018-11-29 · meshnet: mesh neural...

9
MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng, 1 Yifan Feng, 2 Haoxuan You, 1 Xibin Zhao 1* , Yue Gao 1* 1 BNRist, KLISS, School of Software, Tsinghua University, China. 2 School of Information Science and Engineering, Xiamen University {feng-yt15, zxb, gaoyue}@tsinghua.edu.cn, {evanfeng97, haoxuanyou}@gmail.com Abstract Mesh is an important and powerful type of data for 3D shapes and widely studied in the field of computer vision and com- puter graphics. Regarding the task of 3D shape representa- tion, there have been extensive research efforts concentrating on how to represent 3D shapes well using volumetric grid, multi-view and point cloud. However, there is little effort on using mesh data in recent years, due to the complexity and irregularity of mesh data. In this paper, we propose a mesh neural network, named MeshNet, to learn 3D shape represen- tation from mesh data. In this method, face-unit and feature splitting are introduced, and a general architecture with avail- able and effective blocks are proposed. In this way, MeshNet is able to solve the complexity and irregularity problem of mesh and conduct 3D shape representation well. We have ap- plied the proposed MeshNet method in the applications of 3D shape classification and retrieval. Experimental results and comparisons with the state-of-the-art methods demonstrate that the proposed MeshNet can achieve satisfying 3D shape classification and retrieval performance, which indicates the effectiveness of the proposed method on 3D shape represen- tation. Introduction Three-dimensional (3D) shape representation is one of the most fundamental topics in the field of computer vision and computer graphics. In recent years, with the increasing ap- plications in 3D shapes, extensive efforts (Wu et al. 2015; Chang et al. 2015) have been concentrated on 3D shape rep- resentation and proposed methods are successfully applied for different tasks, such as classification and retrieval. For 3D shapes, there are several popular types of data, in- cluding volumetric grid, multi-view, point cloud and mesh. With the success of deep learning methods in computer vi- sion, many neural network methods have been introduced to conduct 3D shape representation using volumetric grid (Wu et al. 2015; Maturana and Scherer 2015), multi-view (Su et al. 2015) and point cloud (Qi et al. 2017a). PointNet (Qi et al. 2017a) proposes to learn on point cloud directly and solves the disorder problem with per-point Multi-Layer- Perceptron (MLP) and a symmetry function. As shown in * Corresponding authors Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 2003 2015 2016 2017 2018 SPH LFD ShapeNet MVCNN FPNN Pairwise Pointnet Pointnet++ Kd-Net SO-Net GVCNN Mesh Volume View Point Cloud Acc Year 60 70 80 90 Figure 1: The developing history of 3D shape representa- tion using different types of data. The X-axis indicates the proposed time of each method, and the Y-axis indicates the classification accuracy. Figure 1, although there have been recent successful meth- ods using the types of volumetric grid, multi-view and point cloud, for the mesh data, there are only early methods using handcraft features directly, such as the Spherical Harmonic descriptor (SPH) (Kazhdan, Funkhouser, and Rusinkiewicz 2003), which limits the applications of mesh data. Mesh data of 3D shapes is a collection of vertices, edges and faces, which is dominantly used in computer graphics for rendering and storing 3D models. Mesh data has the properties of complexity and irregularity. The complexity problem is that mesh consists of multiple elements, and dif- ferent types of connections may be defined among them. The irregularity is another challenge for mesh data process- ing, which indicates that the number of elements in mesh may vary dramatically among 3D shapes, and permutations of them are arbitrary. In spite of these problems, mesh has stronger ability for 3D shape description than other types of data. Under such circumstances, how to effectively represent 3D shapes using mesh data is an urgent and challenging task. In this paper, we present a mesh neural network, named MeshNet, that learns on mesh data directly for 3D shape representation. To deal with the challenges in mesh data pro- cessing, the faces are regarded as the unit and connections arXiv:1811.11424v1 [cs.CV] 28 Nov 2018

Upload: others

Post on 29-May-2020

11 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

MeshNet: Mesh Neural Network for 3D Shape Representation

Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin Zhao1∗, Yue Gao1∗1BNRist, KLISS, School of Software, Tsinghua University, China.

2School of Information Science and Engineering, Xiamen University{feng-yt15, zxb, gaoyue}@tsinghua.edu.cn, {evanfeng97, haoxuanyou}@gmail.com

Abstract

Mesh is an important and powerful type of data for 3D shapesand widely studied in the field of computer vision and com-puter graphics. Regarding the task of 3D shape representa-tion, there have been extensive research efforts concentratingon how to represent 3D shapes well using volumetric grid,multi-view and point cloud. However, there is little effort onusing mesh data in recent years, due to the complexity andirregularity of mesh data. In this paper, we propose a meshneural network, named MeshNet, to learn 3D shape represen-tation from mesh data. In this method, face-unit and featuresplitting are introduced, and a general architecture with avail-able and effective blocks are proposed. In this way, MeshNetis able to solve the complexity and irregularity problem ofmesh and conduct 3D shape representation well. We have ap-plied the proposed MeshNet method in the applications of 3Dshape classification and retrieval. Experimental results andcomparisons with the state-of-the-art methods demonstratethat the proposed MeshNet can achieve satisfying 3D shapeclassification and retrieval performance, which indicates theeffectiveness of the proposed method on 3D shape represen-tation.

IntroductionThree-dimensional (3D) shape representation is one of themost fundamental topics in the field of computer vision andcomputer graphics. In recent years, with the increasing ap-plications in 3D shapes, extensive efforts (Wu et al. 2015;Chang et al. 2015) have been concentrated on 3D shape rep-resentation and proposed methods are successfully appliedfor different tasks, such as classification and retrieval.

For 3D shapes, there are several popular types of data, in-cluding volumetric grid, multi-view, point cloud and mesh.With the success of deep learning methods in computer vi-sion, many neural network methods have been introducedto conduct 3D shape representation using volumetric grid(Wu et al. 2015; Maturana and Scherer 2015), multi-view(Su et al. 2015) and point cloud (Qi et al. 2017a). PointNet(Qi et al. 2017a) proposes to learn on point cloud directlyand solves the disorder problem with per-point Multi-Layer-Perceptron (MLP) and a symmetry function. As shown in

∗Corresponding authorsCopyright c© 2019, Association for the Advancement of ArtificialIntelligence (www.aaai.org). All rights reserved.

2003 2015 2016 2017 2018

SPH

LFDShapeNet

MVCNNFPNN

PairwisePointnet

Pointnet++Kd-Net

SO-NetGVCNN

Mesh

Volume

ViewPoint Cloud

Acc

Year

60

70

80

90

Figure 1: The developing history of 3D shape representa-tion using different types of data. The X-axis indicates theproposed time of each method, and the Y-axis indicates theclassification accuracy.

Figure 1, although there have been recent successful meth-ods using the types of volumetric grid, multi-view and pointcloud, for the mesh data, there are only early methods usinghandcraft features directly, such as the Spherical Harmonicdescriptor (SPH) (Kazhdan, Funkhouser, and Rusinkiewicz2003), which limits the applications of mesh data.

Mesh data of 3D shapes is a collection of vertices, edgesand faces, which is dominantly used in computer graphicsfor rendering and storing 3D models. Mesh data has theproperties of complexity and irregularity. The complexityproblem is that mesh consists of multiple elements, and dif-ferent types of connections may be defined among them.The irregularity is another challenge for mesh data process-ing, which indicates that the number of elements in meshmay vary dramatically among 3D shapes, and permutationsof them are arbitrary. In spite of these problems, mesh hasstronger ability for 3D shape description than other types ofdata. Under such circumstances, how to effectively represent3D shapes using mesh data is an urgent and challenging task.

In this paper, we present a mesh neural network, namedMeshNet, that learns on mesh data directly for 3D shaperepresentation. To deal with the challenges in mesh data pro-cessing, the faces are regarded as the unit and connections

arX

iv:1

811.

1142

4v1

[cs

.CV

] 2

8 N

ov 2

018

Page 2: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

n x

3 MLP (64, 64)

n x

64

Spatial Descriptor

Center

n x

9 Face RotateConvolution

n x

3

Face Kernel Correlationn

x 3

n x

131

n x

64n

x 64

n x

3

MLP(131�����)

NeignborIndex

Normal

Corner

Structural Descriptor

MLP(1024)

Corner

n x

64

MeshConv

n x

131

n x

256

n x

256

MeshConv

n x

512

n x

512

n x

512

n x

512

n x

1024Spatial

Descriptor

StructuralDescriptor

Output Scores

n x

512

n x

256 MLP

(1024)

1024

Global Feature

MLP

(512

, 256

, k)

n x

9n

x 3

n x

3

n x 3NeighborIndex

Normal

CenterPooling

Figure 2: The architecture of MeshNet. The input is a list of faces with initial values, which are fed into the spatial andstructural descriptors to generate initial spatial and structural features. The features are then aggregated with neighboring in-formation in the mesh convolution blocks labeled as “Mesh Conv”, and fed into a pooling function to output the global featureused for further tasks. Multi-layer-perceptron is labeled as “mlp” with the numbers in the parentheses indicating the dimensionof the hidden layers and output layer.

between faces sharing common edges are defined, whichenables us to solve the complexity and irregularity problemwith per-face processes and a symmetry function. Moreover,the feature of faces is split into spatial and structural fea-tures. Based on these ideas, we design the network architec-ture, with two blocks named spatial and structural descrip-tors for learning the initial features, and a mesh convolutionblock for aggregating neighboring features. In this way, theproposed method is able to solve the complexity and irregu-larity problem of mesh and represent 3D shapes well.

We apply our MeshNet method in the tasks of 3D shapeclassification and retrieval on the ModelNet40 (Wu et al.2015) dataset. And the experimental results show that Mesh-Net achieve significant improvement on 3D shape classifi-cation and retrieval using mesh data and comparable perfor-mance with recent methods using other types of 3D data.

The key contributions of our work are as follows:• We propose a neural network using mesh for 3D shape

representation and design blocks for capturing and aggre-gating features of polygon faces in 3D shapes.

• We conduct extensive experiments to evaluate the perfor-mance of the proposed method, and the experimental re-sults show that the proposed method performs well on the3D shape classification and retrieval task.

Related WorkMesh Feature ExtractionThere are plenty of handcraft descriptors that extract fea-tures from mesh. Lien and Kajiya calculate moments of eachtetrahedron in mesh (Lien and Kajiya 1984), and Zhang and

Chen develop more functions applied to each triangle andadd all the resulting values as features (Zhang and Chen2001). Hubeli and Gross extend the features of surfaces toa multiresolution setting to solve the unstructured problemof mesh data (Hubeli and Gross 2001). In SPH (Kazhdan,Funkhouser, and Rusinkiewicz 2003), a rotation invariantrepresentation is presented with existing orientation depen-dent descriptors. Mesh difference of Gaussians (DOG) intro-duces the Gaussian filtering to shape functions.(Zaharescu etal. 2009) Intrinsic shape context (ISC) descriptor (Kokkinoset al. 2012) develops a generalization to surfaces and solvesthe problem of orientational ambiguity.

Deep Learning Methods for 3D ShapeRepresentationWith the construction of large-scale 3D model datasets, nu-merous deep descriptors of 3D shapes are proposed. Basedon different types of data, these methods can be categorizedinto four types.

Voxel-based method. 3DShapeNets (Wu et al. 2015) andVoxNet (Maturana and Scherer 2015) propose to learn onvolumetric grids, which partition the space into regularcubes. However, they introduce extra computation cost dueto the sparsity of data, which restricts them to be applied onmore complex data. Field probing neural networks (FPNN)(Li et al. 2016), Vote3D (Wang and Posner 2015) andOctree-based convolutional neural network (OCNN) (Wanget al. 2017) address the sparsity problem, while they are stillrestricted with input getting larger.

View-based method. Using 2D images of 3D shapes torepresent them is proposed by Multi-view convolutional

Page 3: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

neural networks (MVCNN) (Su et al. 2015), which aggre-gates 2D views from a loop around the object and applies 2Ddeep learning framework to them. Group-view convolutionalneural networks (GVCNN) (Feng et al. 2018) proposes ahierarchical framework, which divides views into differentgroups with different weights to generate a more discrimi-native descriptor for a 3D shape. This type of method alsoexpensively adds the computation cost and is hard to be ap-plied for tasks in larger scenes.

Point-based method. Due to the irregularity of data, pointcloud is not suitable for previous frameworks. PointNet (Qiet al. 2017b) solves this problem with per-point processesand a symmetry function, while it ignores the local informa-tion of points. PointNet++ (Qi et al. 2017b) adds aggregationwith neighbors to solve this problem. Self-organizing net-work (SO-Net) (Li, Chen, and Lee 2018), kernel correlationnetwork (KCNet) (Shen et al. ) and PointSIFT (Jiang, Wu,and Lu 2018) develop more detailed approaches for captur-ing local structures with nearest neighbors. Kd-Net (Klokovand Lempitsky 2017) proposes another approach to solve theirregularity problem using k-d tree.

Fusion method. These methods learn on multiple typesof data and fusion the features of them together. FusionNet(Hegde and Zadeh 2016) uses the volumetric grid and multi-view for classification. Point-view network (PVNet) (You etal. 2018) proposes the embedding attention fusion to exploitboth point cloud data and multi-view data.

MethodIn this block, we present the design of MeshNet. Firstly, weanalyze the properties of mesh, propose the methods for de-signing network and reorganize the input data. We then in-troduce the overall architecture of MeshNet and some blocksfor capturing features of faces and aggregating them withneighbor information, which are then discussed in detail.

Overall Design of MeshNetWe first introduce the mesh data and analyze its properties.Mesh data of 3D shapes is a collection of vertices, edges andfaces, in which vertices are connected with edges and closedsets of edges form faces. In this paper, we only consider tri-angular faces. Mesh data is dominantly used for storing andrendering 3D models in computer graphics, because it pro-vides an approximation of the smooth surfaces of objectsand simplifies the rendering process. Numerous studies on3D shapes in the field of computer graphic and geometricmodeling are taken based on mesh.

Mesh data shows stronger ability to describe 3D shapescomparing with other popular types of data. Volumetric gridand multi-view are data types defined to avoid the irregu-larity of the native data such as mesh and point cloud, whilethey lose some natural information of the original object. Forpoint cloud, there may be ambiguity caused by random sam-pling and the ambiguity is more obvious with fewer amountof points. In contrast, mesh is more clear and loses less nat-ural information. Besides, when capturing local structures,most methods based on point cloud collect the nearest neigh-bors to approximately construct an adjacency matrix for fur-ther process, while in mesh there are explicit connection

NeighborIndex

Normal

Center

Corner

�!OA

<latexit sha1_base64="5uufU2fHqR6H/BMq1AX8F+x2X0A=">AAAC2HicjVHLSsNAFD2Nr1pf1S7dBIvgqiQi6LLqxp0V7ANtkSQd69A0EyYTpZSCO3HrD7jVLxL/QP/CO2MEH4hOSHLm3HvOzL3Xj0OeKMd5zlkTk1PTM/nZwtz8wuJScXmlkYhUBqweiFDIlu8lLOQRqyuuQtaKJfMGfsiafn9fx5uXTCZcRMdqGLPOwOtF/JwHniLqrFhqCwpL3rtQnpTianS4Oz4rlp2KY5b9E7gZKCNbNVF8QhtdCARIMQBDBEU4hIeEnlO4cBAT18GIOEmImzjDGAXSppTFKMMjtk/fHu1OMzaivfZMjDqgU0J6JSltrJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQoXxP6l+8j8r07XonCOHVMDp5piw+jqgswlNV3RN7c/VaXIISZO4y7FJeHAKD/6bBtNYmrXvfVM/MVkalbvgyw3xau+JQ3Y/T7On6CxWXGdinu0Va7uZaPOYxVr2KB5bqOKA9RQJ+8h7vGAR+vEurZurNv3VCuXaUr4sqy7N5wMl9M=</latexit><latexit sha1_base64="5uufU2fHqR6H/BMq1AX8F+x2X0A=">AAAC2HicjVHLSsNAFD2Nr1pf1S7dBIvgqiQi6LLqxp0V7ANtkSQd69A0EyYTpZSCO3HrD7jVLxL/QP/CO2MEH4hOSHLm3HvOzL3Xj0OeKMd5zlkTk1PTM/nZwtz8wuJScXmlkYhUBqweiFDIlu8lLOQRqyuuQtaKJfMGfsiafn9fx5uXTCZcRMdqGLPOwOtF/JwHniLqrFhqCwpL3rtQnpTianS4Oz4rlp2KY5b9E7gZKCNbNVF8QhtdCARIMQBDBEU4hIeEnlO4cBAT18GIOEmImzjDGAXSppTFKMMjtk/fHu1OMzaivfZMjDqgU0J6JSltrJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQoXxP6l+8j8r07XonCOHVMDp5piw+jqgswlNV3RN7c/VaXIISZO4y7FJeHAKD/6bBtNYmrXvfVM/MVkalbvgyw3xau+JQ3Y/T7On6CxWXGdinu0Va7uZaPOYxVr2KB5bqOKA9RQJ+8h7vGAR+vEurZurNv3VCuXaUr4sqy7N5wMl9M=</latexit><latexit sha1_base64="5uufU2fHqR6H/BMq1AX8F+x2X0A=">AAAC2HicjVHLSsNAFD2Nr1pf1S7dBIvgqiQi6LLqxp0V7ANtkSQd69A0EyYTpZSCO3HrD7jVLxL/QP/CO2MEH4hOSHLm3HvOzL3Xj0OeKMd5zlkTk1PTM/nZwtz8wuJScXmlkYhUBqweiFDIlu8lLOQRqyuuQtaKJfMGfsiafn9fx5uXTCZcRMdqGLPOwOtF/JwHniLqrFhqCwpL3rtQnpTianS4Oz4rlp2KY5b9E7gZKCNbNVF8QhtdCARIMQBDBEU4hIeEnlO4cBAT18GIOEmImzjDGAXSppTFKMMjtk/fHu1OMzaivfZMjDqgU0J6JSltrJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQoXxP6l+8j8r07XonCOHVMDp5piw+jqgswlNV3RN7c/VaXIISZO4y7FJeHAKD/6bBtNYmrXvfVM/MVkalbvgyw3xau+JQ3Y/T7On6CxWXGdinu0Va7uZaPOYxVr2KB5bqOKA9RQJ+8h7vGAR+vEurZurNv3VCuXaUr4sqy7N5wMl9M=</latexit><latexit sha1_base64="5uufU2fHqR6H/BMq1AX8F+x2X0A=">AAAC2HicjVHLSsNAFD2Nr1pf1S7dBIvgqiQi6LLqxp0V7ANtkSQd69A0EyYTpZSCO3HrD7jVLxL/QP/CO2MEH4hOSHLm3HvOzL3Xj0OeKMd5zlkTk1PTM/nZwtz8wuJScXmlkYhUBqweiFDIlu8lLOQRqyuuQtaKJfMGfsiafn9fx5uXTCZcRMdqGLPOwOtF/JwHniLqrFhqCwpL3rtQnpTianS4Oz4rlp2KY5b9E7gZKCNbNVF8QhtdCARIMQBDBEU4hIeEnlO4cBAT18GIOEmImzjDGAXSppTFKMMjtk/fHu1OMzaivfZMjDqgU0J6JSltrJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQoXxP6l+8j8r07XonCOHVMDp5piw+jqgswlNV3RN7c/VaXIISZO4y7FJeHAKD/6bBtNYmrXvfVM/MVkalbvgyw3xau+JQ3Y/T7On6CxWXGdinu0Va7uZaPOYxVr2KB5bqOKA9RQJ+8h7vGAR+vEurZurNv3VCuXaUr4sqy7N5wMl9M=</latexit>��!OB

<latexit sha1_base64="G2RBOz3IuuhRjXaHO0sr/u5gxI8=">AAAC2HicjVHLSsNAFD2Nr/qOdukmWARXJRFBl0U37qxgH1iLJHGsQ9NMmEyUUgR34tYfcKtfJP6B/oV3xhTUIjohyZlz7zkz994giXiqXPe1YE1MTk3PFGfn5hcWl5btldVGKjIZsnooIiFbgZ+yiMesrriKWCuRzO8HEWsGvX0db14xmXIRH6tBwjp9vxvzCx76iqgzu3QqKCx591L5Uorr4eHezZlddiuuWc448HJQRr5qwn7BKc4hECJDHwwxFOEIPlJ62vDgIiGugyFxkhA3cYYbzJE2oyxGGT6xPfp2adfO2Zj22jM16pBOieiVpHSwQRpBeZKwPs0x8cw4a/Y376Hx1Hcb0D/IvfrEKlwS+5dulPlfna5F4QK7pgZONSWG0dWFuUtmuqJv7nypSpFDQpzG5xSXhEOjHPXZMZrU1K5765v4m8nUrN6HeW6Gd31LGrD3c5zjoLFV8dyKd7Rdru7loy5iDevYpHnuoIoD1FAn7wEe8YRn68S6te6s+89Uq5BrSvi2rIcPnm2X1A==</latexit><latexit sha1_base64="G2RBOz3IuuhRjXaHO0sr/u5gxI8=">AAAC2HicjVHLSsNAFD2Nr/qOdukmWARXJRFBl0U37qxgH1iLJHGsQ9NMmEyUUgR34tYfcKtfJP6B/oV3xhTUIjohyZlz7zkz994giXiqXPe1YE1MTk3PFGfn5hcWl5btldVGKjIZsnooIiFbgZ+yiMesrriKWCuRzO8HEWsGvX0db14xmXIRH6tBwjp9vxvzCx76iqgzu3QqKCx591L5Uorr4eHezZlddiuuWc448HJQRr5qwn7BKc4hECJDHwwxFOEIPlJ62vDgIiGugyFxkhA3cYYbzJE2oyxGGT6xPfp2adfO2Zj22jM16pBOieiVpHSwQRpBeZKwPs0x8cw4a/Y376Hx1Hcb0D/IvfrEKlwS+5dulPlfna5F4QK7pgZONSWG0dWFuUtmuqJv7nypSpFDQpzG5xSXhEOjHPXZMZrU1K5765v4m8nUrN6HeW6Gd31LGrD3c5zjoLFV8dyKd7Rdru7loy5iDevYpHnuoIoD1FAn7wEe8YRn68S6te6s+89Uq5BrSvi2rIcPnm2X1A==</latexit><latexit sha1_base64="G2RBOz3IuuhRjXaHO0sr/u5gxI8=">AAAC2HicjVHLSsNAFD2Nr/qOdukmWARXJRFBl0U37qxgH1iLJHGsQ9NMmEyUUgR34tYfcKtfJP6B/oV3xhTUIjohyZlz7zkz994giXiqXPe1YE1MTk3PFGfn5hcWl5btldVGKjIZsnooIiFbgZ+yiMesrriKWCuRzO8HEWsGvX0db14xmXIRH6tBwjp9vxvzCx76iqgzu3QqKCx591L5Uorr4eHezZlddiuuWc448HJQRr5qwn7BKc4hECJDHwwxFOEIPlJ62vDgIiGugyFxkhA3cYYbzJE2oyxGGT6xPfp2adfO2Zj22jM16pBOieiVpHSwQRpBeZKwPs0x8cw4a/Y376Hx1Hcb0D/IvfrEKlwS+5dulPlfna5F4QK7pgZONSWG0dWFuUtmuqJv7nypSpFDQpzG5xSXhEOjHPXZMZrU1K5765v4m8nUrN6HeW6Gd31LGrD3c5zjoLFV8dyKd7Rdru7loy5iDevYpHnuoIoD1FAn7wEe8YRn68S6te6s+89Uq5BrSvi2rIcPnm2X1A==</latexit><latexit sha1_base64="G2RBOz3IuuhRjXaHO0sr/u5gxI8=">AAAC2HicjVHLSsNAFD2Nr/qOdukmWARXJRFBl0U37qxgH1iLJHGsQ9NMmEyUUgR34tYfcKtfJP6B/oV3xhTUIjohyZlz7zkz994giXiqXPe1YE1MTk3PFGfn5hcWl5btldVGKjIZsnooIiFbgZ+yiMesrriKWCuRzO8HEWsGvX0db14xmXIRH6tBwjp9vxvzCx76iqgzu3QqKCx591L5Uorr4eHezZlddiuuWc448HJQRr5qwn7BKc4hECJDHwwxFOEIPlJ62vDgIiGugyFxkhA3cYYbzJE2oyxGGT6xPfp2adfO2Zj22jM16pBOieiVpHSwQRpBeZKwPs0x8cw4a/Y376Hx1Hcb0D/IvfrEKlwS+5dulPlfna5F4QK7pgZONSWG0dWFuUtmuqJv7nypSpFDQpzG5xSXhEOjHPXZMZrU1K5765v4m8nUrN6HeW6Gd31LGrD3c5zjoLFV8dyKd7Rdru7loy5iDevYpHnuoIoD1FAn7wEe8YRn68S6te6s+89Uq5BrSvi2rIcPnm2X1A==</latexit>��!OC

<latexit sha1_base64="B2raW6ovXDf/LKQ1DZtXNdGAz4E=">AAAC2HicjVHLSsNAFD3GV62vqks3wSK4KokIuix2484K1hZbKUk6bQfTTJhMlFIK7sStP+BWv0j8A/0L74wR1CI6IcmZc+85M/dePw55ohznZcqanpmdm88t5BeXlldWC2vrZ4lIZcBqgQiFbPhewkIesZriKmSNWDJv4Ies7l9WdLx+xWTCRXSqhjG7GHi9iHd54Cmi2oWNlqCw5L2+8qQU16PjyrhdKDolxyx7ErgZKCJbVVF4RgsdCARIMQBDBEU4hIeEniZcOIiJu8CIOEmImzjDGHnSppTFKMMj9pK+Pdo1MzaivfZMjDqgU0J6JSltbJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQp9Yv/SfWb+V6drUejiwNTAqabYMLq6IHNJTVf0ze0vVSlyiInTuENxSTgwys8+20aTmNp1bz0TfzWZmtX7IMtN8aZvSQN2f45zEpztllyn5J7sFcuH2ahz2MQWdmie+yjjCFXUyHuIBzziyTq3bqxb6+4j1ZrKNBv4tqz7d6DOl9U=</latexit><latexit sha1_base64="B2raW6ovXDf/LKQ1DZtXNdGAz4E=">AAAC2HicjVHLSsNAFD3GV62vqks3wSK4KokIuix2484K1hZbKUk6bQfTTJhMlFIK7sStP+BWv0j8A/0L74wR1CI6IcmZc+85M/dePw55ohznZcqanpmdm88t5BeXlldWC2vrZ4lIZcBqgQiFbPhewkIesZriKmSNWDJv4Ies7l9WdLx+xWTCRXSqhjG7GHi9iHd54Cmi2oWNlqCw5L2+8qQU16PjyrhdKDolxyx7ErgZKCJbVVF4RgsdCARIMQBDBEU4hIeEniZcOIiJu8CIOEmImzjDGHnSppTFKMMj9pK+Pdo1MzaivfZMjDqgU0J6JSltbJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQp9Yv/SfWb+V6drUejiwNTAqabYMLq6IHNJTVf0ze0vVSlyiInTuENxSTgwys8+20aTmNp1bz0TfzWZmtX7IMtN8aZvSQN2f45zEpztllyn5J7sFcuH2ahz2MQWdmie+yjjCFXUyHuIBzziyTq3bqxb6+4j1ZrKNBv4tqz7d6DOl9U=</latexit><latexit sha1_base64="B2raW6ovXDf/LKQ1DZtXNdGAz4E=">AAAC2HicjVHLSsNAFD3GV62vqks3wSK4KokIuix2484K1hZbKUk6bQfTTJhMlFIK7sStP+BWv0j8A/0L74wR1CI6IcmZc+85M/dePw55ohznZcqanpmdm88t5BeXlldWC2vrZ4lIZcBqgQiFbPhewkIesZriKmSNWDJv4Ies7l9WdLx+xWTCRXSqhjG7GHi9iHd54Cmi2oWNlqCw5L2+8qQU16PjyrhdKDolxyx7ErgZKCJbVVF4RgsdCARIMQBDBEU4hIeEniZcOIiJu8CIOEmImzjDGHnSppTFKMMj9pK+Pdo1MzaivfZMjDqgU0J6JSltbJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQp9Yv/SfWb+V6drUejiwNTAqabYMLq6IHNJTVf0ze0vVSlyiInTuENxSTgwys8+20aTmNp1bz0TfzWZmtX7IMtN8aZvSQN2f45zEpztllyn5J7sFcuH2ahz2MQWdmie+yjjCFXUyHuIBzziyTq3bqxb6+4j1ZrKNBv4tqz7d6DOl9U=</latexit><latexit sha1_base64="B2raW6ovXDf/LKQ1DZtXNdGAz4E=">AAAC2HicjVHLSsNAFD3GV62vqks3wSK4KokIuix2484K1hZbKUk6bQfTTJhMlFIK7sStP+BWv0j8A/0L74wR1CI6IcmZc+85M/dePw55ohznZcqanpmdm88t5BeXlldWC2vrZ4lIZcBqgQiFbPhewkIesZriKmSNWDJv4Ies7l9WdLx+xWTCRXSqhjG7GHi9iHd54Cmi2oWNlqCw5L2+8qQU16PjyrhdKDolxyx7ErgZKCJbVVF4RgsdCARIMQBDBEU4hIeEniZcOIiJu8CIOEmImzjDGHnSppTFKMMj9pK+Pdo1MzaivfZMjDqgU0J6JSltbJNGUJ4krE+zTTw1zpr9zXtkPPXdhvT3M68BsQp9Yv/SfWb+V6drUejiwNTAqabYMLq6IHNJTVf0ze0vVSlyiInTuENxSTgwys8+20aTmNp1bz0TfzWZmtX7IMtN8aZvSQN2f45zEpztllyn5J7sFcuH2ahz2MQWdmie+yjjCFXUyHuIBzziyTq3bqxb6+4j1ZrKNBv4tqz7d6DOl9U=</latexit>

�!n<latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit>

O<latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit>

�!n<latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit><latexit sha1_base64="clgDmGxq+gdQeCFlTTEsGoQwqic=">AAAC13icjVHLSsNAFD3GV62vWpdugkVwVRIRdCm6cVnBakWLJONYB9NMmEx8UIo7cesPuNU/Ev9A/8I74xR8IDohyZlz7zkz9944S0Sug+BlyBseGR0bL02UJ6emZ2Yrc9W9XBaK8SaTiVStOMp5IlLe1EInvJUpHnXjhO/H51smvn/BVS5kuquvM97uRp1UnAoWaaKOK9UjSWElOmc6Ukpe9tL+caUW1AO7/J8gdKAGtxqy8owjnECCoUAXHCk04QQRcnoOESJARlwbPeIUIWHjHH2USVtQFqeMiNhz+nZod+jYlPbGM7dqRqck9CpS+lgijaQ8Rdic5tt4YZ0N+5t3z3qau13TP3ZeXWI1zoj9SzfI/K/O1KJxinVbg6CaMsuY6phzKWxXzM39T1VpcsiIM/iE4oows8pBn32ryW3tpreRjb/aTMOaPXO5Bd7MLWnA4fdx/gR7K/UwqIc7q7WNTTfqEhawiGWa5xo2sI0GmuR9hQc84sk78G68W+/uI9Ubcpp5fFne/Tsd/pen</latexit>

d<latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit>

e<latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit>

f<latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit>

O<latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit>n faces

A<latexit sha1_base64="1fpNIUZRPl2vZJVNl9EHXpCLbCI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkVxGUL9gFaJEmndeg0CZOJUIr+gFv9NvEP9C+8M05BLaITkpw5954zc+8NU8Ez5XmvBWdufmFxqbhcWlldW98ob261siSXEWtGiUhkJwwyJnjMmoorwTqpZMEoFKwdDs90vH3HZMaT+FKNU9YdBYOY93kUKKIaJzflilf1zHJngW9BBXbVk/ILrtFDggg5RmCIoQgLBMjouYIPDylxXUyIk4S4iTPco0TanLIYZQTEDuk7oN2VZWPaa8/MqCM6RdArSelijzQJ5UnC+jTXxHPjrNnfvCfGU99tTP/Qeo2IVbgl9i/dNPO/Ol2LQh/HpgZONaWG0dVF1iU3XdE3d79UpcghJU7jHsUl4cgop312jSYzteveBib+ZjI1q/eRzc3xrm9JA/Z/jnMWtA6qvlf1G4eV2qkddRE72MU+zfMINVygjqbxfsQTnp1zRziZk3+mOgWr2ca35Tx8AOhtj0Y=</latexit><latexit sha1_base64="1fpNIUZRPl2vZJVNl9EHXpCLbCI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkVxGUL9gFaJEmndeg0CZOJUIr+gFv9NvEP9C+8M05BLaITkpw5954zc+8NU8Ez5XmvBWdufmFxqbhcWlldW98ob261siSXEWtGiUhkJwwyJnjMmoorwTqpZMEoFKwdDs90vH3HZMaT+FKNU9YdBYOY93kUKKIaJzflilf1zHJngW9BBXbVk/ILrtFDggg5RmCIoQgLBMjouYIPDylxXUyIk4S4iTPco0TanLIYZQTEDuk7oN2VZWPaa8/MqCM6RdArSelijzQJ5UnC+jTXxHPjrNnfvCfGU99tTP/Qeo2IVbgl9i/dNPO/Ol2LQh/HpgZONaWG0dVF1iU3XdE3d79UpcghJU7jHsUl4cgop312jSYzteveBib+ZjI1q/eRzc3xrm9JA/Z/jnMWtA6qvlf1G4eV2qkddRE72MU+zfMINVygjqbxfsQTnp1zRziZk3+mOgWr2ca35Tx8AOhtj0Y=</latexit><latexit sha1_base64="1fpNIUZRPl2vZJVNl9EHXpCLbCI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkVxGUL9gFaJEmndeg0CZOJUIr+gFv9NvEP9C+8M05BLaITkpw5954zc+8NU8Ez5XmvBWdufmFxqbhcWlldW98ob261siSXEWtGiUhkJwwyJnjMmoorwTqpZMEoFKwdDs90vH3HZMaT+FKNU9YdBYOY93kUKKIaJzflilf1zHJngW9BBXbVk/ILrtFDggg5RmCIoQgLBMjouYIPDylxXUyIk4S4iTPco0TanLIYZQTEDuk7oN2VZWPaa8/MqCM6RdArSelijzQJ5UnC+jTXxHPjrNnfvCfGU99tTP/Qeo2IVbgl9i/dNPO/Ol2LQh/HpgZONaWG0dVF1iU3XdE3d79UpcghJU7jHsUl4cgop312jSYzteveBib+ZjI1q/eRzc3xrm9JA/Z/jnMWtA6qvlf1G4eV2qkddRE72MU+zfMINVygjqbxfsQTnp1zRziZk3+mOgWr2ca35Tx8AOhtj0Y=</latexit><latexit sha1_base64="1fpNIUZRPl2vZJVNl9EHXpCLbCI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkVxGUL9gFaJEmndeg0CZOJUIr+gFv9NvEP9C+8M05BLaITkpw5954zc+8NU8Ez5XmvBWdufmFxqbhcWlldW98ob261siSXEWtGiUhkJwwyJnjMmoorwTqpZMEoFKwdDs90vH3HZMaT+FKNU9YdBYOY93kUKKIaJzflilf1zHJngW9BBXbVk/ILrtFDggg5RmCIoQgLBMjouYIPDylxXUyIk4S4iTPco0TanLIYZQTEDuk7oN2VZWPaa8/MqCM6RdArSelijzQJ5UnC+jTXxHPjrNnfvCfGU99tTP/Qeo2IVbgl9i/dNPO/Ol2LQh/HpgZONaWG0dVF1iU3XdE3d79UpcghJU7jHsUl4cgop312jSYzteveBib+ZjI1q/eRzc3xrm9JA/Z/jnMWtA6qvlf1G4eV2qkddRE72MU+zfMINVygjqbxfsQTnp1zRziZk3+mOgWr2ca35Tx8AOhtj0Y=</latexit>

B<latexit sha1_base64="OB+MD7FAoYbDCJJnnCYjDlaEKJo=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy1JBXLZgH1CLJNNpDZ0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKqJMed5rwVlYXFpeKa6urW9sbm2XdnZbWZJLxpssEYnshEHGRRTzpoqU4J1U8mAcCt4OR+c63r7jMouS+EpNUt4bB8M4GkQsUEQ1ajelslfxzHLngW9BGXbVk9ILrtFHAoYcY3DEUIQFAmT0dOHDQ0pcD1PiJKHIxDnusUbanLI4ZQTEjug7pF3XsjHttWdm1IxOEfRKUro4JE1CeZKwPs018dw4a/Y376nx1Heb0D+0XmNiFW6J/Us3y/yvTteiMMCZqSGimlLD6OqYdclNV/TN3S9VKXJIidO4T3FJmBnlrM+u0WSmdt3bwMTfTKZm9Z7Z3Bzv+pY0YP/nOOdB67jiexW/cVKu1uyoi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED6s2PRw==</latexit><latexit sha1_base64="OB+MD7FAoYbDCJJnnCYjDlaEKJo=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy1JBXLZgH1CLJNNpDZ0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKqJMed5rwVlYXFpeKa6urW9sbm2XdnZbWZJLxpssEYnshEHGRRTzpoqU4J1U8mAcCt4OR+c63r7jMouS+EpNUt4bB8M4GkQsUEQ1ajelslfxzHLngW9BGXbVk9ILrtFHAoYcY3DEUIQFAmT0dOHDQ0pcD1PiJKHIxDnusUbanLI4ZQTEjug7pF3XsjHttWdm1IxOEfRKUro4JE1CeZKwPs018dw4a/Y376nx1Heb0D+0XmNiFW6J/Us3y/yvTteiMMCZqSGimlLD6OqYdclNV/TN3S9VKXJIidO4T3FJmBnlrM+u0WSmdt3bwMTfTKZm9Z7Z3Bzv+pY0YP/nOOdB67jiexW/cVKu1uyoi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED6s2PRw==</latexit><latexit sha1_base64="OB+MD7FAoYbDCJJnnCYjDlaEKJo=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy1JBXLZgH1CLJNNpDZ0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKqJMed5rwVlYXFpeKa6urW9sbm2XdnZbWZJLxpssEYnshEHGRRTzpoqU4J1U8mAcCt4OR+c63r7jMouS+EpNUt4bB8M4GkQsUEQ1ajelslfxzHLngW9BGXbVk9ILrtFHAoYcY3DEUIQFAmT0dOHDQ0pcD1PiJKHIxDnusUbanLI4ZQTEjug7pF3XsjHttWdm1IxOEfRKUro4JE1CeZKwPs018dw4a/Y376nx1Heb0D+0XmNiFW6J/Us3y/yvTteiMMCZqSGimlLD6OqYdclNV/TN3S9VKXJIidO4T3FJmBnlrM+u0WSmdt3bwMTfTKZm9Z7Z3Bzv+pY0YP/nOOdB67jiexW/cVKu1uyoi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED6s2PRw==</latexit><latexit sha1_base64="OB+MD7FAoYbDCJJnnCYjDlaEKJo=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy1JBXLZgH1CLJNNpDZ0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKqJMed5rwVlYXFpeKa6urW9sbm2XdnZbWZJLxpssEYnshEHGRRTzpoqU4J1U8mAcCt4OR+c63r7jMouS+EpNUt4bB8M4GkQsUEQ1ajelslfxzHLngW9BGXbVk9ILrtFHAoYcY3DEUIQFAmT0dOHDQ0pcD1PiJKHIxDnusUbanLI4ZQTEjug7pF3XsjHttWdm1IxOEfRKUro4JE1CeZKwPs018dw4a/Y376nx1Heb0D+0XmNiFW6J/Us3y/yvTteiMMCZqSGimlLD6OqYdclNV/TN3S9VKXJIidO4T3FJmBnlrM+u0WSmdt3bwMTfTKZm9Z7Z3Bzv+pY0YP/nOOdB67jiexW/cVKu1uyoi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED6s2PRw==</latexit>

C<latexit sha1_base64="9JnBjP+aj5VVLWDvCKMvwNY+8Zk=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy2JBXLZgH1CLJOm0Dp0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKnimPO+14CwsLi2vFFfX1jc2t7ZLO7utLMllxJpRIhLZCYOMCR6zpuJKsE4qWTAOBWuHo5qOt++YzHgSX6lJynrjYBjzAY8CRVSjdlMqexXPLHce+BaUYVc9Kb3gGn0kiJBjDIYYirBAgIyeLnx4SInrYUqcJMRNnOEea6TNKYtRRkDsiL5D2nUtG9Nee2ZGHdEpgl5JSheHpEkoTxLWp7kmnhtnzf7mPTWe+m4T+ofWa0yswi2xf+lmmf/V6VoUBjgzNXCqKTWMri6yLrnpir65+6UqRQ4pcRr3KS4JR0Y567NrNJmpXfc2MPE3k6lZvY9sbo53fUsasP9znPOgdVzxvYrfOClXz+2oi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED7S2PSA==</latexit><latexit sha1_base64="9JnBjP+aj5VVLWDvCKMvwNY+8Zk=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy2JBXLZgH1CLJOm0Dp0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKnimPO+14CwsLi2vFFfX1jc2t7ZLO7utLMllxJpRIhLZCYOMCR6zpuJKsE4qWTAOBWuHo5qOt++YzHgSX6lJynrjYBjzAY8CRVSjdlMqexXPLHce+BaUYVc9Kb3gGn0kiJBjDIYYirBAgIyeLnx4SInrYUqcJMRNnOEea6TNKYtRRkDsiL5D2nUtG9Nee2ZGHdEpgl5JSheHpEkoTxLWp7kmnhtnzf7mPTWe+m4T+ofWa0yswi2xf+lmmf/V6VoUBjgzNXCqKTWMri6yLrnpir65+6UqRQ4pcRr3KS4JR0Y567NrNJmpXfc2MPE3k6lZvY9sbo53fUsasP9znPOgdVzxvYrfOClXz+2oi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED7S2PSA==</latexit><latexit sha1_base64="9JnBjP+aj5VVLWDvCKMvwNY+8Zk=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy2JBXLZgH1CLJOm0Dp0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKnimPO+14CwsLi2vFFfX1jc2t7ZLO7utLMllxJpRIhLZCYOMCR6zpuJKsE4qWTAOBWuHo5qOt++YzHgSX6lJynrjYBjzAY8CRVSjdlMqexXPLHce+BaUYVc9Kb3gGn0kiJBjDIYYirBAgIyeLnx4SInrYUqcJMRNnOEea6TNKYtRRkDsiL5D2nUtG9Nee2ZGHdEpgl5JSheHpEkoTxLWp7kmnhtnzf7mPTWe+m4T+ofWa0yswi2xf+lmmf/V6VoUBjgzNXCqKTWMri6yLrnpir65+6UqRQ4pcRr3KS4JR0Y567NrNJmpXfc2MPE3k6lZvY9sbo53fUsasP9znPOgdVzxvYrfOClXz+2oi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED7S2PSA==</latexit><latexit sha1_base64="9JnBjP+aj5VVLWDvCKMvwNY+8Zk=">AAACxHicjVHLSsNAFD2Nr/quunQTLIKrkoigy2JBXLZgH1CLJOm0Dp0mYTIRStEfcKvfJv6B/oV3ximoRXRCkjPn3nNm7r1hKnimPO+14CwsLi2vFFfX1jc2t7ZLO7utLMllxJpRIhLZCYOMCR6zpuJKsE4qWTAOBWuHo5qOt++YzHgSX6lJynrjYBjzAY8CRVSjdlMqexXPLHce+BaUYVc9Kb3gGn0kiJBjDIYYirBAgIyeLnx4SInrYUqcJMRNnOEea6TNKYtRRkDsiL5D2nUtG9Nee2ZGHdEpgl5JSheHpEkoTxLWp7kmnhtnzf7mPTWe+m4T+ofWa0yswi2xf+lmmf/V6VoUBjgzNXCqKTWMri6yLrnpir65+6UqRQ4pcRr3KS4JR0Y567NrNJmpXfc2MPE3k6lZvY9sbo53fUsasP9znPOgdVzxvYrfOClXz+2oi9jHAY5onqeo4hJ1NI33I57w7Fw4wsmc/DPVKVjNHr4t5+ED7S2PSA==</latexit>

O<latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit><latexit sha1_base64="O5qDU1NnK9phbsF6iL+vaKQAXO0=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxJ0t2AdokSSd1qHTJEwmQin6A27128Q/0L/wzjgFtYhOSHLm3HvOzL03TAXPlOe9Fpy5+YXFpeJyaWV1bX2jvLnVypJcRqwZJSKRnTDImOAxayquBOukkgWjULB2ODzV8fYdkxlP4ks1Tll3FAxi3udRoIhqXNyUK17VM8udBb4FFdhVT8ovuEYPCSLkGIEhhiIsECCj5wo+PKTEdTEhThLiJs5wjxJpc8pilBEQO6TvgHZXlo1prz0zo47oFEGvJKWLPdIklCcJ69NcE8+Ns2Z/854YT323Mf1D6zUiVuGW2L9008z/6nQtCn0cmxo41ZQaRlcXWZfcdEXf3P1SlSKHlDiNexSXhCOjnPbZNZrM1K57G5j4m8nUrN5HNjfHu74lDdj/Oc5Z0Dqo+l7VbxxWaid21EXsYBf7NM8j1HCOOprG+xFPeHbOHOFkTv6Z6hSsZhvflvPwAQm8j1Q=</latexit>

n x 3

n x 9

n x 3

n x 3

0.3

0.4

0.2

1 x 3

0.2

0.7

0.1

-0.2

1 x 9

1.0

0.0

0.0

1 x 3

34

78

83

1 x 3

one face

d<latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit><latexit sha1_base64="7kVD1bztflqXqy6ue1pPjTm2YK4=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SJJOa+g0CTMToRT9Abf6beIf6F94Z5yCWkQnJDlz7j1n5t4bZjyWyvNeC87C4tLySnG1tLa+sblV3t5pyTQXEWtGKU9FJwwk43HCmipWnHUywYJxyFk7HJ3rePuOCRmnyZWaZKw3DoZJPIijQBHV6N+UK17VM8udB74FFdhVT8svuEYfKSLkGIMhgSLMEUDS04UPDxlxPUyJE4RiE2e4R4m0OWUxygiIHdF3SLuuZRPaa09p1BGdwukVpHRxQJqU8gRhfZpr4rlx1uxv3lPjqe82oX9ovcbEKtwS+5dulvlfna5FYYBTU0NMNWWG0dVF1iU3XdE3d79UpcghI07jPsUF4cgoZ312jUaa2nVvAxN/M5ma1fvI5uZ417ekAfs/xzkPWkdV36v6jeNK7cyOuog97OOQ5nmCGi5RR9N4P+IJz86Fwx3p5J+pTsFqdvFtOQ8fO5yPaQ==</latexit>

e<latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit><latexit sha1_base64="mJ8GqMi7I82JglfZ7jvtceK19vA=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qFpEmYmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bppFQ2vNeC87C4tLySnG1tLa+sblV3t5pqSSTjDdZEiWyEwaKRyLmTS10xDup5ME4jHg7HJ2bePuOSyWS+EpPUt4bB8NYDAQLNFENflOueFXPLnce+DmoIF/1pPyCa/SRgCHDGBwxNOEIARQ9XfjwkBLXw5Q4SUjYOMc9SqTNKItTRkDsiL5D2nVzNqa98VRWzeiUiF5JShcHpEkoTxI2p7k2nllnw/7mPbWe5m4T+oe515hYjVti/9LNMv+rM7VoDHBqaxBUU2oZUx3LXTLbFXNz90tVmhxS4gzuU1wSZlY567NrNcrWbnob2PibzTSs2bM8N8O7uSUN2P85znnQOqr6XtVvHFdqZ/moi9jDPg5pnieo4RJ1NK33I57w7Fw4kaOc7DPVKeSaXXxbzsMHPfyPag==</latexit>

f<latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit><latexit sha1_base64="7QQfQx4iNX04F8go9Nvv8s3hbAI=">AAACxHicjVHLSsNAFD2Nr1pfVZdugkVwVRIRdFkUxGUL9gG1SDKd1qHTJGQmQin6A27128Q/0L/wzpiCWkQnJDlz7j1n5t4bJlIo7XmvBWdhcWl5pbhaWlvf2Nwqb++0VJyljDdZLOO0EwaKSxHxphZa8k6S8mAcSt4OR+cm3r7jqRJxdKUnCe+Ng2EkBoIFmqjG4KZc8aqeXe488HNQQb7qcfkF1+gjBkOGMTgiaMISARQ9XfjwkBDXw5S4lJCwcY57lEibURanjIDYEX2HtOvmbER746msmtEpkt6UlC4OSBNTXkrYnObaeGadDfub99R6mrtN6B/mXmNiNW6J/Us3y/yvztSiMcCprUFQTYllTHUsd8lsV8zN3S9VaXJIiDO4T/GUMLPKWZ9dq1G2dtPbwMbfbKZhzZ7luRnezS1pwP7Pcc6D1lHV96p+47hSO8tHXcQe9nFI8zxBDZeoo2m9H/GEZ+fCkY5yss9Up5BrdvFtOQ8fQFyPaw==</latexit>

Figure 3: Initial values of each face. There are four typesof initial values, divided into two parts: center, corner andnormal are the face information, and neighbor index is theneighbor information.

relationships to show the local structure clearly. However,mesh data is also more irregular and complex for the multi-ple compositions and varying numbers of elements.

To get full use of the advantages of mesh and solve theproblem of its irregularity and complexity, we propose twokey ideas of design:

• Regard face as the unit. Mesh data consists of multipleelements and connections may be defined among them.To simplify the data organization, we regard face as theonly unit and define a connection between two faces ifthey share a common edge. There are several advantagesof this simplification. First is that one triangular face canconnect with no more than three faces, which makes theconnection relationship regular and easy to use. More im-portantly, we can solve the disorder problem with per-faceprocesses and a symmetry function, which is similar toPointNet (Qi et al. 2017a), with per-face processes anda symmetry function. And intuitively, face also containsmore information than vertex and edge.

• Split feature of face. Though the above simplificationenables us to consume mesh data similar to point-basedmethods, there are still some differences between point-unit and face-unit because face contains more informationthan point. We only need to know “where you are” for apoint, while we also want to know “what you look like”for a face. Correspondingly, we split the feature of facesinto spatial feature and structural feature, which helpsus to capture features more explicitly.

Following the above ideas, we transform the mesh datainto a list of faces. For each face, we define the initial valuesof it, which are divided into two parts (illustrated in Fig 3):

• Face Information:

Page 4: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

– Center: coordinate of the center point– Corner: vectors from the center point to three vertices– Normal: the unit normal vector

• Neighbor Information:– Neighbor Index: indexes of the connected faces (filled

with the index of itself if the face connects with lessthan three faces)

In the final of this section, we present the overall archi-tecture of MeshNet, illustrated in Fig 2. A list of faces withinitial values is fed into two blocks, named spatial descrip-tor and structural descriptor, to generate the initial spatialand structural features of faces. The features are then passedthrough some mesh convolution blocks to aggregate neigh-boring information, which get features of two types as inputand output new features of them. It is noted that all the pro-cesses above work on each face respectively and share thesame parameters. After these processes, a pooling functionis applied to features of all faces for generating global fea-ture, which is used for further tasks. The above blocks willbe discussed in following sections.

Spatial and Structural DescriptorsWe split feature of faces into spatial feature and structuralfeature. The spatial feature is expected to be relevant to thespatial position of faces, and the structural feature is rele-vant to the shape information and local structures. In thissection, we present the design of two blocks, named spatialand structural descriptors, for generating initial spatial andstructural features.

Spatial descriptor The only input value relevant to spatialposition is the center value. In this block, we simply applya shared MLP to each face’s center, similar to the methodsbased on point cloud, and output initial spatial feature.

Structural descriptor: face rotate convolution We pro-pose two types of structural descriptor, and the first one isnamed face rotate convolution, which captures the “inner”structure of faces and focus on the shape information offaces. The input of this block is the corner value.

The operation of this block is illustrated in Fig 4. Supposethe corner vectors of a face are v1,v2,v3, we define theoutput value of this block as follows:

g(1

3(f(v1,v2) + f(v2,v3) + f(v3,v1))), (1)

where f(·, ·) : R3 × R3 → RK1 and g(·) : RK1 → RK2 areany valid functions.

This process is similar to a convolution operation, withtwo vectors as the kernel size, one vector as the stride andK1

as the number of kernels, except that translation of kernelsis replaced by rotation. The kernels, represented by f(·, ·),rotates through the face and works on two vectors each time.

With the above process, we eliminate the influence causedby the order of processing corners, avoid individually con-sidering each corner and also leave full space for mining

v1<latexit sha1_base64="oY9xBoyCVPHzsCLD8qgLPTDP65g=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQi6LLoxmUF+4CmlMl00g6dTMLMTaGE/oYbF4q49Wfc+TdO2iy09cDA4Zx7uWdOkEhh0HW/ndLG5tb2Tnm3srd/cHhUPT5pmzjVjLdYLGPdDajhUijeQoGSdxPNaRRI3gkm97nfmXJtRKyecJbwfkRHSoSCUbSS70cUx0GYTecDb1CtuXV3AbJOvILUoEBzUP3yhzFLI66QSWpMz3MT7GdUo2CSzyt+anhC2YSOeM9SRSNu+tki85xcWGVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqaKLcFb/fI6aV/VPbfuPV7XGndFHWU4g3O4BA9uoAEP0IQWMEjgGV7hzUmdF+fd+ViOlpxi5xT+wPn8ASdSkb4=</latexit><latexit sha1_base64="oY9xBoyCVPHzsCLD8qgLPTDP65g=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQi6LLoxmUF+4CmlMl00g6dTMLMTaGE/oYbF4q49Wfc+TdO2iy09cDA4Zx7uWdOkEhh0HW/ndLG5tb2Tnm3srd/cHhUPT5pmzjVjLdYLGPdDajhUijeQoGSdxPNaRRI3gkm97nfmXJtRKyecJbwfkRHSoSCUbSS70cUx0GYTecDb1CtuXV3AbJOvILUoEBzUP3yhzFLI66QSWpMz3MT7GdUo2CSzyt+anhC2YSOeM9SRSNu+tki85xcWGVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqaKLcFb/fI6aV/VPbfuPV7XGndFHWU4g3O4BA9uoAEP0IQWMEjgGV7hzUmdF+fd+ViOlpxi5xT+wPn8ASdSkb4=</latexit><latexit sha1_base64="oY9xBoyCVPHzsCLD8qgLPTDP65g=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQi6LLoxmUF+4CmlMl00g6dTMLMTaGE/oYbF4q49Wfc+TdO2iy09cDA4Zx7uWdOkEhh0HW/ndLG5tb2Tnm3srd/cHhUPT5pmzjVjLdYLGPdDajhUijeQoGSdxPNaRRI3gkm97nfmXJtRKyecJbwfkRHSoSCUbSS70cUx0GYTecDb1CtuXV3AbJOvILUoEBzUP3yhzFLI66QSWpMz3MT7GdUo2CSzyt+anhC2YSOeM9SRSNu+tki85xcWGVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqaKLcFb/fI6aV/VPbfuPV7XGndFHWU4g3O4BA9uoAEP0IQWMEjgGV7hzUmdF+fd+ViOlpxi5xT+wPn8ASdSkb4=</latexit><latexit sha1_base64="oY9xBoyCVPHzsCLD8qgLPTDP65g=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQi6LLoxmUF+4CmlMl00g6dTMLMTaGE/oYbF4q49Wfc+TdO2iy09cDA4Zx7uWdOkEhh0HW/ndLG5tb2Tnm3srd/cHhUPT5pmzjVjLdYLGPdDajhUijeQoGSdxPNaRRI3gkm97nfmXJtRKyecJbwfkRHSoSCUbSS70cUx0GYTecDb1CtuXV3AbJOvILUoEBzUP3yhzFLI66QSWpMz3MT7GdUo2CSzyt+anhC2YSOeM9SRSNu+tki85xcWGVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqaKLcFb/fI6aV/VPbfuPV7XGndFHWU4g3O4BA9uoAEP0IQWMEjgGV7hzUmdF+fd+ViOlpxi5xT+wPn8ASdSkb4=</latexit> v2

<latexit sha1_base64="ssjW1CGTywUuhiz3AoMjfAjFPK8=">AAAB83icbVDLSsNAFL3xWeur6tLNYBFclaQIuiy6cVnBPqApZTKdtEMnkzBzUyihv+HGhSJu/Rl3/o2TNgttPTBwOOde7pkTJFIYdN1vZ2Nza3tnt7RX3j84PDqunJy2TZxqxlsslrHuBtRwKRRvoUDJu4nmNAok7wST+9zvTLk2IlZPOEt4P6IjJULBKFrJ9yOK4yDMpvNBfVCpujV3AbJOvIJUoUBzUPnyhzFLI66QSWpMz3MT7GdUo2CSz8t+anhC2YSOeM9SRSNu+tki85xcWmVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqayLcFb/fI6addrnlvzHq+rjbuijhKcwwVcgQc30IAHaEILGCTwDK/w5qTOi/PufCxHN5xi5wz+wPn8ASjWkb8=</latexit><latexit sha1_base64="ssjW1CGTywUuhiz3AoMjfAjFPK8=">AAAB83icbVDLSsNAFL3xWeur6tLNYBFclaQIuiy6cVnBPqApZTKdtEMnkzBzUyihv+HGhSJu/Rl3/o2TNgttPTBwOOde7pkTJFIYdN1vZ2Nza3tnt7RX3j84PDqunJy2TZxqxlsslrHuBtRwKRRvoUDJu4nmNAok7wST+9zvTLk2IlZPOEt4P6IjJULBKFrJ9yOK4yDMpvNBfVCpujV3AbJOvIJUoUBzUPnyhzFLI66QSWpMz3MT7GdUo2CSz8t+anhC2YSOeM9SRSNu+tki85xcWmVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqayLcFb/fI6addrnlvzHq+rjbuijhKcwwVcgQc30IAHaEILGCTwDK/w5qTOi/PufCxHN5xi5wz+wPn8ASjWkb8=</latexit><latexit sha1_base64="ssjW1CGTywUuhiz3AoMjfAjFPK8=">AAAB83icbVDLSsNAFL3xWeur6tLNYBFclaQIuiy6cVnBPqApZTKdtEMnkzBzUyihv+HGhSJu/Rl3/o2TNgttPTBwOOde7pkTJFIYdN1vZ2Nza3tnt7RX3j84PDqunJy2TZxqxlsslrHuBtRwKRRvoUDJu4nmNAok7wST+9zvTLk2IlZPOEt4P6IjJULBKFrJ9yOK4yDMpvNBfVCpujV3AbJOvIJUoUBzUPnyhzFLI66QSWpMz3MT7GdUo2CSz8t+anhC2YSOeM9SRSNu+tki85xcWmVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqayLcFb/fI6addrnlvzHq+rjbuijhKcwwVcgQc30IAHaEILGCTwDK/w5qTOi/PufCxHN5xi5wz+wPn8ASjWkb8=</latexit><latexit sha1_base64="ssjW1CGTywUuhiz3AoMjfAjFPK8=">AAAB83icbVDLSsNAFL3xWeur6tLNYBFclaQIuiy6cVnBPqApZTKdtEMnkzBzUyihv+HGhSJu/Rl3/o2TNgttPTBwOOde7pkTJFIYdN1vZ2Nza3tnt7RX3j84PDqunJy2TZxqxlsslrHuBtRwKRRvoUDJu4nmNAok7wST+9zvTLk2IlZPOEt4P6IjJULBKFrJ9yOK4yDMpvNBfVCpujV3AbJOvIJUoUBzUPnyhzFLI66QSWpMz3MT7GdUo2CSz8t+anhC2YSOeM9SRSNu+tki85xcWmVIwljbp5As1N8bGY2MmUWBncwzmlUvF//zeimGt/1MqCRFrtjyUJhKgjHJCyBDoTlDObOEMi1sVsLGVFOGtqayLcFb/fI6addrnlvzHq+rjbuijhKcwwVcgQc30IAHaEILGCTwDK/w5qTOi/PufCxHN5xi5wz+wPn8ASjWkb8=</latexit>

v3<latexit sha1_base64="jpc+PIP/B+GyVY2jx67zQgO78ZI=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQq6LLoxmUF+4CmlMn0ph06mYSZSaGE/oYbF4q49Wfc+TdO2iy0emDgcM693DMnSATXxnW/nNLa+sbmVnm7srO7t39QPTxq6zhVDFssFrHqBlSj4BJbhhuB3UQhjQKBnWByl/udKSrNY/loZgn2IzqSPOSMGiv5fkTNOAiz6XxwOajW3Lq7APlLvILUoEBzUP30hzFLI5SGCap1z3MT08+oMpwJnFf8VGNC2YSOsGeppBHqfrbIPCdnVhmSMFb2SUMW6s+NjEZaz6LATuYZ9aqXi/95vdSEN/2MyyQ1KNnyUJgKYmKSF0CGXCEzYmYJZYrbrISNqaLM2JoqtgRv9ct/Sfui7rl17+Gq1rgt6ijDCZzCOXhwDQ24hya0gEECT/ACr07qPDtvzvtytOQUO8fwC87HNypakcA=</latexit><latexit sha1_base64="jpc+PIP/B+GyVY2jx67zQgO78ZI=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQq6LLoxmUF+4CmlMn0ph06mYSZSaGE/oYbF4q49Wfc+TdO2iy0emDgcM693DMnSATXxnW/nNLa+sbmVnm7srO7t39QPTxq6zhVDFssFrHqBlSj4BJbhhuB3UQhjQKBnWByl/udKSrNY/loZgn2IzqSPOSMGiv5fkTNOAiz6XxwOajW3Lq7APlLvILUoEBzUP30hzFLI5SGCap1z3MT08+oMpwJnFf8VGNC2YSOsGeppBHqfrbIPCdnVhmSMFb2SUMW6s+NjEZaz6LATuYZ9aqXi/95vdSEN/2MyyQ1KNnyUJgKYmKSF0CGXCEzYmYJZYrbrISNqaLM2JoqtgRv9ct/Sfui7rl17+Gq1rgt6ijDCZzCOXhwDQ24hya0gEECT/ACr07qPDtvzvtytOQUO8fwC87HNypakcA=</latexit><latexit sha1_base64="jpc+PIP/B+GyVY2jx67zQgO78ZI=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQq6LLoxmUF+4CmlMn0ph06mYSZSaGE/oYbF4q49Wfc+TdO2iy0emDgcM693DMnSATXxnW/nNLa+sbmVnm7srO7t39QPTxq6zhVDFssFrHqBlSj4BJbhhuB3UQhjQKBnWByl/udKSrNY/loZgn2IzqSPOSMGiv5fkTNOAiz6XxwOajW3Lq7APlLvILUoEBzUP30hzFLI5SGCap1z3MT08+oMpwJnFf8VGNC2YSOsGeppBHqfrbIPCdnVhmSMFb2SUMW6s+NjEZaz6LATuYZ9aqXi/95vdSEN/2MyyQ1KNnyUJgKYmKSF0CGXCEzYmYJZYrbrISNqaLM2JoqtgRv9ct/Sfui7rl17+Gq1rgt6ijDCZzCOXhwDQ24hya0gEECT/ACr07qPDtvzvtytOQUO8fwC87HNypakcA=</latexit><latexit sha1_base64="jpc+PIP/B+GyVY2jx67zQgO78ZI=">AAAB83icbVDLSsNAFL2pr1pfVZduBovgqiQq6LLoxmUF+4CmlMn0ph06mYSZSaGE/oYbF4q49Wfc+TdO2iy0emDgcM693DMnSATXxnW/nNLa+sbmVnm7srO7t39QPTxq6zhVDFssFrHqBlSj4BJbhhuB3UQhjQKBnWByl/udKSrNY/loZgn2IzqSPOSMGiv5fkTNOAiz6XxwOajW3Lq7APlLvILUoEBzUP30hzFLI5SGCap1z3MT08+oMpwJnFf8VGNC2YSOsGeppBHqfrbIPCdnVhmSMFb2SUMW6s+NjEZaz6LATuYZ9aqXi/95vdSEN/2MyyQ1KNnyUJgKYmKSF0CGXCEzYmYJZYrbrISNqaLM2JoqtgRv9ct/Sfui7rl17+Gq1rgt6ijDCZzCOXhwDQ24hya0gEECT/ACr07qPDtvzvtytOQUO8fwC87HNypakcA=</latexit>

⇥<latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit>

⇥<latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit>

⇥<latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit>

Kernel

Kernel

Kernel

Pooling

MLP

⇥<latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> <latexit sha1_base64="EKJy6rjXhiAKrBUXGSg3J2VqHFQ=">AAAB7XicbVDLSgNBEOyNrxhfUY9eBoPgKeyKoMegF48RzAOSJcxOZpMxszPLTK8QQv7BiwdFvPo/3vwbJ8keNLGgoajqprsrSqWw6PvfXmFtfWNzq7hd2tnd2z8oHx41rc4M4w2mpTbtiFouheINFCh5OzWcJpHkrWh0O/NbT9xYodUDjlMeJnSgRCwYRSc1uygSbnvlil/15yCrJMhJBXLUe+Wvbl+zLOEKmaTWdgI/xXBCDQom+bTUzSxPKRvRAe84qqhbEk7m107JmVP6JNbGlUIyV39PTGhi7TiJXGdCcWiXvZn4n9fJML4OJ0KlGXLFFoviTBLUZPY66QvDGcqxI5QZ4W4lbEgNZegCKrkQguWXV0nzohr41eD+slK7yeMowgmcwjkEcAU1uIM6NIDBIzzDK7x52nvx3r2PRWvBy2eO4Q+8zx+2K480</latexit> : Convolution

shared

shared

Figure 4: The face rotate convolution block. Kernels rotatethrough the face and are applied to pairs of corner vectors forthe convolution operation.

features inside faces. After the rotate convolution, we applyan average pooling and a shared MLP as g(·) to each face,and output features with the length of K2.

Structural descriptor: face kernel correlation Anotherstructural descriptor we design is the face kernel correla-tion, aiming to capture the “outer” structure of faces andexplore the environments where faces locate. The methodis inspired by KCNet (Shen et al. ), which uses kernel corre-lation (KC) (Tsin and Kanade 2004) for mining local struc-tures in point clouds. KCNet learns kernels representing dif-ferent spatial distributions of point sets, and measures thegeometric affinities between kernels and neighboring pointsfor each point to indicate the local structure. However, thismethod is also restricted by the ambiguity of point cloud,and may achieve better performance in mesh.

In our face kernel correlation, we select the normal valuesof each face and its neighbors as the source, and learnablesets of vectors as the reference kernels. Since all the normalswe use are unit vectors, we model vectors of kernels with pa-rameters in the spherical coordinate system, and parameters(θ, φ) represent the unit vector (x, y, z) in the Euclidean co-ordinate system: {

x = sin θ cosφy = sin θ sinφz = cos θ

, (2)

where θ ∈ [0, π] and φ ∈ [0, 2π).We define the kernel correlation between the i-th face and

the k-th kernel as follows:

KC(i, k) =1

|Ni||Mk|∑n∈Ni

∑m∈Mk

Kσ(n,m), (3)

where Ni is the set of normals of the i-th face and its neigh-bor faces, Mk is the set of normals in the k-th kernel, and

Page 5: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

Kσ(·, ·) : R3 × R3 → R is the kernel function indicatingthe affinity between two vectors. In this paper, we generallychoose the Gaussian kernel:

Kσ(n,m) = exp(−‖n−m‖22σ2

), (4)

where ‖·‖ is the length of a vector in the Euclidean space,and σ is the hyper-parameter that controls the kernels’ re-solving ability or tolerance to the varying of sources.

With the above definition, we calculate the kernel corre-lation between each face and kernel, and more similar pairswill get higher values. Since the parameters of kernels arelearnable, they will turn to some common distributions onthe surfaces of 3D shapes and be able to describe the localstructures of faces. We set the value of KC(i, k) as the k-thfeature of the i-th face. Therefore, withM learnable kernels,we generate features with the length of M for each face.

Mesh ConvolutionThe mesh convolution block is designed to expand the re-ceptive field of faces, which denotes the number of facesperceived by each face, by aggregating information of neigh-boring faces. In this process, features related to spatial po-sitions should not be included directly because we focus onfaces in a local area and should not be influenced by wherethe area locates. In the 2D convolutional neural network,both the convolution and pooling operations do not intro-duce any positional information directly while aggregatingwith neighboring pixels’ features. Since we have taken outthe structural feature that is irrelevant to positions, we onlyaggregate them in this block.

Aggregation of structural feature enables us to capturestructures of wider field around each face. Furthermore, toget more comprehensive feature, we also combine the spa-tial and structural feature together in this block. The meshconvolution block contains two parts: combination of spa-tial and structural features and aggregation of structuralfeature, which respectively output the new spatial and struc-tural features. Fig 5 illustrates the design of this block.

Combination of Spatial and Structural Features Weuse one of the most common methods of combining twotypes of features, which concatenates them together and ap-plies a MLP. As we have mentioned, the combination result,as the new spatial feature, is actually more comprehensiveand contains both spatial and structural information. There-fore, in the pipeline of our network, we concatenate all thecombination results for generating global feature.

Aggregation of Structural Feature With the input struc-tural feature and neighbor index, we aggregate the feature ofa face with features of its connected faces. Several aggrega-tion methods are listed and discussed as follows:• Average pooling: The average pooling may be the most

common aggregation method, which simply calculates theaverage value of features in each channel. However, thismethod sometimes weakens the strongly-activated areasof the original features and reduce the distinction of them.

NeighborFeatures

n⇥in

2n⇥3

SelfFeature

ŏŏ n

⇥ou

t 2

Aggregation

MLP

(out2)

Gathering

max pooling

max pooling

Aggregation for one face

MLP

average pooling

1

2 3

Concatenation1Max pooling2Average pooling3

Combination

AggregationNeighbor

Index

SpatialFeature

StructuralFeature

n⇥in

1n⇥in

2 n⇥

out 1

n⇥ou

t 2

n⇥3 n⇥in

1n⇥in

2

n⇥in

1

n⇥

in2

n⇥

out 1

MLP

(out1)

Combination

Figure 5: The mesh convolution. “Combination” donatesthe combination of spatial and structural feature. “Aggrega-tion” denotes the aggregation of structural feature, in which“Gathering” denotes the process of getting neighbors’ fea-tures and “Aggregation for one face” denotes different meth-ods of aggregating features.

• Max pooling: Another pooling method is the max pool-ing, which calculates the max value of features in eachchannel. Max pooling is widely used in 2D and 3D deeplearning frameworks for its advantage of maintaining thestrong activation of neurons.

• Concatenation: We define the concatenation aggrega-tion, which concatenates the feature of face with featureof each neighbor respectively, passes these pairs througha shared MLP and applies a max pooling to the results.This method both keeps the original activation and leavesspace for the network to combine neighboring features.We finally use the concatenation method in this paper. Af-

ter aggregation, another MLP is applied to further fusion theneighboring features and output new structural feature.

Implementation DetailsNow we present the details of implementing MeshNet, il-lustrated in Fig 2, including the settings of hyper-parametersand some details of the overall architecture.

The spatial descriptor contains fully-connected layers (64,64) and output a initial spatial feature with length of 64. Pa-rameters inside parentheses indicate the dimensions of lay-ers except the input layer. In the face rotate convolution, weset K1 = 32 and K2 = 64, and correspondingly, the func-tions f(·, ·) and g(·) are implemented by fully-connectedlayers (32, 32) and (64, 64). In the face kernel correlation,we set M = 64 (64 kernels) and σ = 0.2.

We parameterize the mesh convolution block with a four-tuple (in1, in2, out1, out2), where in1 and out1 indicate theinput and output channels of spatial feature, and the in2 andout2 indicates the same of structural feature. The two mesh

Page 6: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

convolution blocks used in the pipeline of MeshNet are con-figured as (64, 131, 256, 256) and (256, 256, 512, 512).

ExperimentsIn the experiments, we first apply our network for 3D shapeclassification and retrieval. Then we conduct detailed abla-tion experiments to analyze the effectiveness of blocks in thearchitecture. We also investigate the robustness to the num-ber of faces and the time and space complexity of our net-work. Finally, we visualize the structural features from thetwo structural descriptors.

3D Shape Classification and RetrievalWe apply our network on the ModelNet40 dataset (Wu etal. 2015) for classification and retrieval tasks. The datasetcontains 12,311 mesh models from 40 categories, in which9,843 models for training and 2,468 models for testing. Foreach model, we simplify the mesh data into no more than1,024 faces, translate it to the geometric center, and nor-malize it into a unit sphere. Moreover, we also compute thenormal vector and indexes of connected faces for each face.During the training, we augment the data by jittering the po-sitions of vertices by a Gaussian noise with zero mean and0.01 standard deviation. Since the number of faces variesamong models, we randomly fill the list of faces to the lengthof 1024 with existing faces for batch training.

For classification, we apply fully-connected layers (512,256, 40) to the global features as the classifier, and adddropout layers with drop probability of 0.5 before the lasttwo fully-connected layers. For retrieval, we calculate theL2 distances between the global features as similarities andevaluate the result with mean average precision (mAP). Weuse the SGD optimizer for training, with initial learning rate0.01, momentum 0.9, weight decay 0.0005 and batch size64.

Method Modality Acc mAP(%) (%)

3DShapeNets (Wu et al. 2015) volume 77.3 49.2VoxNet (Maturana and Scherer2015)

volume 83.0 -

FPNN (Li et al. 2016) volume 88.4 -

LFD (Chen et al. 2003) view 75.5 40.9MVCNN (Su et al. 2015) view 90.1 79.5Pairwise (Johns, Leutenegger,and Davison 2016)

view 90.7 -

PointNet (Qi et al. 2017a) point 89.2 -PointNet++ (Qi et al. 2017b) point 90.7 -Kd-Net (Klokov and Lempit-sky 2017)

point 91.8 -

KCNet (Shen et al. ) point 91.0 -

SPH (Kazhdan, Funkhouser,and Rusinkiewicz 2003)

mesh 68.2 33.3

MeshNet mesh 91.9 81.9

Table 1: Classification and retrieval results on ModelNet40.

Spatial X X X X XStructural-FRC X X X XStructural-FKC X X X XMesh Conv X X X X X

Accuracy (%) 83.5 88.2 87.0 89.9 90.4 91.9

Table 2: Classification results of ablation experiments onModelNet40.

Aggregation Method Accuracy (%)

Average Pooling 90.7Max Pooling 91.1Concatenation 91.9

Table 3: Classification results of different aggregation meth-ods on ModelNet40.

Table 1 shows the experimental results of classificationand retrieval on ModelNet40, comparing our work with rep-resentative methods. It is shown that, as a mesh-based rep-resentation, MeshNet achieves satisfying performance andmakes great improvement compared with traditional mesh-based methods. It is also comparable with recent deep learn-ing methods based on other types of data.

Our performance is dedicated to the following reasons.With face-unit and per-face processes, MeshNet solves thecomplexity and irregularity problem of mesh data and makesit suitable for deep learning method. Though with deeplearning’s strong ability to capture features, we do not sim-ply apply it, but design blocks to get full use of the rich infor-mation in mesh. Splitting features into spatial and structuralfeatures enables us to consider the spatial distribution andlocal structure of shapes respectively. And the mesh convo-lution blocks widen the receptive field of faces. Therefore,the proposed method is able to capture detailed features offaces and conduct the 3D shape representation well.

On the Effectiveness of BlocksTo analyze the design of blocks in our architecture and provethe effectiveness of them, we conduct several ablation exper-iments, which compare the classification results while vary-ing the settings of architecture or removing some blocks.

For the spatial descriptor, labeled as “Spatial” in Table 2,we remove it together with the use of spatial feature in thenetwork, and maintain the aggregation of structural featurein the mesh convolution.

For the structural descriptor, we first remove the wholeof it and use max pooling to aggregate the spatial featurein the mesh convolution. Then we partly remove the facerotate convolution, labeled as “Structural-FRC” in Table 2,or the face kernel correlation, labeled as “Structural-FKC”,and keep the rest of pipeline to prove the effectiveness ofeach structural descriptor.

For the mesh convolution, labeled as “Mesh Conv” in Ta-ble 2, we remove it and use the initial features to generatethe global feature directly. We also explore the effectiveness

Page 7: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

of different aggregation methods in this block, and com-pare them in Table 3. The experimental results show that theadopted concatenation method performs better for aggregat-ing neighboring features.

On the Number of FacesThe number of faces in ModelNet40 varies dramaticallyamong models. To explore the robustness of MeshNet to thenumber of faces, we regroup the test data by the number offaces with interval 200. In Table 4, we list the proportion ofthe number of models in each group, together with the clas-sification results. It is shown that the accuracy is absolutelyirrelevant to the number of faces and shows no downtrendwith the decrease of it, which proves the great robustness ofMeshNet to the number of faces. Specifically, on the 9 mod-els with less than 50 faces (the minimum is 10), our networkachieves 100% classification accuracy, showing the abilityto represent models with extremely few faces.

Number of Faces Proportion (%) Accuracy (%)

[1000, 1024) 69.48 92.00[800, 1000) 6.90 92.35[600, 800) 4.70 93.10[400, 600) 6.90 91.76[200, 400) 6.17 90.79[0, 200) 5.84 90.97

Table 4: Classification results of groups with different num-ber of faces on ModelNet40.

On the Time and Space ComplexityTable 5 compares the time and space complexity of our net-work with several representative methods based on othertypes of data for the classification task. The column labeled“#params” shows the total number of parameters in the net-work and the column labeled “FLOPs/sample” shows thenumber of floating-operations conducted for each input sam-ple, representing the space and time complexity respectively.

It is known that methods based on volumetric grid andmulti-view introduce extra computation cost while methodsbased on point cloud are more efficient. Theoretically, ourmethod works with per-face processes and has a linear com-plexity to the number of faces. In Table 5, MeshNet showscomparable effectiveness with methods based on point cloudin both time and space complexity, leaving enough space forfurther development.

Method #params FLOPs /(M) sample (M)

PointNet (Qi et al. 2017a) 3.5 440Subvolume (Qi et al. 2016) 16.6 3633MVCNN (Su et al. 2015) 60.0 62057

MeshNet 4.25 509

Table 5: Time and space complexity for classification.

Figure 6: Feature visualization of structural feature.Models from the same column are colored with their val-ues of the same channel in features. Left: Features from theface rotate convolution. Right: Features from the face kernelcorrelation.

Feature VisualizationTo figure out whether the structural descriptors successfullycapture our features of faces as expected, we visualize thetwo types of structural features from face rotate convolu-tion and face kernel correlation. We randomly choose sev-eral channels of these features, and for each channel, wepaint faces with colors of different depth corresponding totheir values in this channel.

The left of Fig 6 visualizes features from face rotate con-volution, which is expected to capture the “inner” featuresof faces and concern about the shapes of them. It is clearlyshown that these faces with similar look are also coloredsimilarly, and different channels may be activated by differ-ent types of triangular faces.

The visualization results of features from face kernel cor-relation are in the right of Fig 6. As we have mentioned, thisdescriptor captures the “outer” features of each face and isrelevant to the whole appearance of the area where the facelocates. In the visualization, faces in similar types of areas,such as flat surfaces and steep slopes, turn to have similarfeatures, regardless of their own shapes and sizes.

ConclusionsIn this paper, we propose a mesh neural network, namedMeshNet, which learns on mesh data directly for 3D shaperepresentation. The proposed method is able to solve thecomplexity and irregularity problem of mesh data and con-duct 3D shape representation well. In this method, the poly-gon faces are regarded as the unit and features of themare split into spatial and structural features. We also de-sign blocks for capturing and aggregating features of faces.We conduct experiments for 3D shape classification andretrieval and compare our method with the state-of-the-artmethods. The experimental result and comparisons demon-strate the effectiveness of the proposed method on 3D shaperepresentation. In the future, the network can be further de-veloped for more computer vision tasks.

Page 8: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

AcknowledgmentsThis work was supported by National Key R&D Pro-gram of China (Grant No. 2017YFC0113000), NationalNatural Science Funds of China (U1701262, 61671267),National Science and Technology Major Project (No.2016ZX01038101), MIIT IT funds (Research and applica-tion of TCN key technologies) of China, and The NationalKey Technology R&D Program (No. 2015BAG14B01-02).

References[Chang et al. 2015] Chang, A. X.; Funkhouser, T.; Guibas,L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva,M.; Song, S.; Su, H.; et al. 2015. ShapeNet: AnInformation-Rich 3D Model Repository. arXiv preprintarXiv:1512.03012.

[Chen et al. 2003] Chen, D.-Y.; Tian, X.-P.; Shen, Y.-T.; andOuhyoung, M. 2003. On Visual Similarity Based 3D ModelRetrieval. In Computer Graphics Forum, volume 22, 223–232. Wiley Online Library.

[Feng et al. 2018] Feng, Y.; Zhang, Z.; Zhao, X.; Ji, R.; andGao, Y. 2018. GVCNN: Group-View Convolutional NeuralNetworks for 3D Shape Recognition. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion, 264–272.

[Hegde and Zadeh 2016] Hegde, V., and Zadeh, R. 2016. Fu-sionnet: 3D Object Classification Using Multiple Data Rep-resentations. arXiv preprint arXiv:1607.05695.

[Hubeli and Gross 2001] Hubeli, A., and Gross, M. 2001.Multiresolution Feature Extraction for UnstructuredMeshes. In Proceedings of the Conference on Visualization,287–294. IEEE Computer Society.

[Jiang, Wu, and Lu 2018] Jiang, M.; Wu, Y.; and Lu, C.2018. PointSIFT: A SIFT-like Network Module for3D Point Cloud Semantic Segmentation. arXiv preprintarXiv:1807.00652.

[Johns, Leutenegger, and Davison 2016] Johns, E.;Leutenegger, S.; and Davison, A. J. 2016. PairwiseDecomposition of Image Sequences for Active Multi-viewRecognition. In Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, 3813–3822.

[Kazhdan, Funkhouser, and Rusinkiewicz 2003] Kazhdan,M.; Funkhouser, T.; and Rusinkiewicz, S. 2003. RotationInvariant Spherical Harmonic Representation of 3D ShapeDescriptors. In Symposium on Geometry Processing,volume 6, 156–164.

[Klokov and Lempitsky 2017] Klokov, R., and Lempitsky, V.2017. Escape from Cells: Deep Kd-Networks for the Recog-nition of 3D Point Cloud Models. In Proceedings of theIEEE International Conference on Computer Vision, 863–872. IEEE.

[Kokkinos et al. 2012] Kokkinos, I.; Bronstein, M. M.; Lit-man, R.; and Bronstein, A. M. 2012. Intrinsic Shape ContextDescriptors for Deformable Shapes. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion, 159–166. IEEE.

[Li et al. 2016] Li, Y.; Pirk, S.; Su, H.; Qi, C. R.; and Guibas,L. J. 2016. FPNN: Field Probing Neural Networks for 3DData. In Advances in Neural Information Processing Sys-tems, 307–315.

[Li, Chen, and Lee 2018] Li, J.; Chen, B. M.; and Lee, G. H.2018. SO-Net: Self-Organizing Network for Point CloudAnalysis. In Proceedings of the IEEE Conference on Com-puter Vision and Pattern Recognition, 9397–9406.

[Lien and Kajiya 1984] Lien, S.-l., and Kajiya, J. T. 1984. ASymbolic Method for Calculating the Integral Properties ofArbitrary Nonconvex Polyhedra. IEEE Computer Graphicsand Applications 4(10):35–42.

[Maturana and Scherer 2015] Maturana, D., and Scherer, S.2015. VoxNet: A 3D Convolutional Neural Network forReal-time Object Recognition. In 2015 IEEE/RSJ Interna-tional Conference on Intelligent Robots and Systems (IROS),922–928. IEEE.

[Qi et al. 2016] Qi, C. R.; Su, H.; Nießner, M.; Dai, A.; Yan,M.; and Guibas, L. J. 2016. Volumetric and Multi-ViewCNNs for Object Classification on 3D Data. In Proceed-ings of the IEEE Conference on Computer Vision and Pat-tern Recognition, 5648–5656.

[Qi et al. 2017a] Qi, C. R.; Su, H.; Mo, K.; and Guibas, L. J.2017a. PointNet: Deep Learning on Point Sets for 3D Clas-sification and Segmentation. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition,77–85.

[Qi et al. 2017b] Qi, C. R.; Yi, L.; Su, H.; and Guibas, L. J.2017b. PointNet++: Deep Hierarchical Feature Learning onPoint Sets in a Metric Space. In Advances in Neural Infor-mation Processing Systems, 5099–5108.

[Shen et al. ] Shen, Y.; Feng, C.; Yang, Y.; and Tian, D. Min-ing Point Cloud Local Structures by Kernel Correlation andGraph Pooling. In Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, volume 4.

[Su et al. 2015] Su, H.; Maji, S.; Kalogerakis, E.; andLearned-Miller, E. 2015. Multi-View Convolutional NeuralNetworks for 3D Shape Recognition. In Proceedings of theIEEE International Conference on Computer Vision, 945–953.

[Tsin and Kanade 2004] Tsin, Y., and Kanade, T. 2004. ACorrelation-Based Approach to Robust Point Set Registra-tion. In European Conference on Computer Vision, 558–569. Springer.

[Wang and Posner 2015] Wang, D. Z., and Posner, I. 2015.Voting for Voting in Online Point Cloud Object Detection.In Robotics: Science and Systems, volume 1.

[Wang et al. 2017] Wang, P.-S.; Liu, Y.; Guo, Y.-X.; Sun, C.-Y.; and Tong, X. 2017. O-CNN: Octree-based ConvolutionalNeural Networks for 3D Shape Analysis. ACM Transactionson Graphics (TOG) 36(4):72.

[Wu et al. 2015] Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang,L.; Tang, X.; and Xiao, J. 2015. 3d ShapeNets: A DeepRepresentation for Volumetric Shapes. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion, 1912–1920.

Page 9: MeshNet: Mesh Neural Network for 3D Shape Representation · 2018-11-29 · MeshNet: Mesh Neural Network for 3D Shape Representation Yutong Feng,1 Yifan Feng,2 Haoxuan You,1 Xibin

[You et al. 2018] You, H.; Feng, Y.; Ji, R.; and Gao, Y. 2018.PVnet: A Joint Convolutional Network of Point Cloud andMulti-view for 3D Shape Recognition. In Proceedings of the26th ACM International Conference on Multimedia, 1310–1318. ACM.

[Zaharescu et al. 2009] Zaharescu, A.; Boyer, E.; Varanasi,K.; and Horaud, R. 2009. Surface Feature Detection andDescription with Applications to Mesh Matching. In Pro-ceedings of the IEEE Conference on Computer Vision andPattern Recognition, 373–380. IEEE.

[Zhang and Chen 2001] Zhang, C., and Chen, T. 2001. Effi-cient Feature Extraction for 2D/3D Objects in Mesh Repre-sentation. In Proceedings of the IEEE International Confer-ence on Image Processing, volume 3, 935–938. IEEE.