journal of la fsd-10: a dataset for competitive sports content analysis · 2020. 2. 11. · sports...

9
JOURNAL OF L A T E X CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 1 FSD-10: A Dataset for Competitive Sports Content Analysis Shenglan Liu Member, IEEE, Xiang Liu, Gao Huang, Lin Feng, Lianyu Hu, Dong Jiang, Aibin Zhang, Yang Liu, Hong Qiao Fellow, IEEE Abstract—Action recognition is an important and challeng- ing problem in video analysis. Although the past decade has witnessed progress in action recognition with the development of deep learning, such process has been slow in competitive sports content analysis. To promote the research on action recognition from competitive sports video clips, we introduce a Figure Skating Dataset (FSD-10) for finegrained sports content analysis. To this end, we collect 1484 clips from the worldwide figure skating championships in 2017-2018, which consist of 10 different actions in men/ladies programs. Each clip is at a rate of 30 frames per second with resolution 1080 × 720. These clips are then annotated by experts in type, grade of execution, skater info, .etc. To build a baseline for action recognition in figure skating, we evaluate state-of-the-art action recognition methods on FSD-10. Motivated by the idea that domain knowledge is of great concern in sports field, we propose a keyframe based temporal segment network (KTSN) for classification and achieve remarkable performance. Experimental results demonstrate that FSD-10 is an ideal dataset for benchmarking action recognition algorithms, as it requires to accurately extract action motions rather than action poses. We hope FSD-10, which is designed to have a large collection of finegrained actions, can serve as a new challenge to develop more robust and advanced action recognition models. I. I NTRODUCTION Due to the popularity of media-sharing platforms(e.g. YouTube), sports content analysis (SCA[31]) has become an important research topic in computer vision [39][10][30]. A vast amount of sports videos are piled up in computer storage, which are potential resources for deep learning. In recent years, many enterprises (e.g. Bloomberg, SAP) have focus on SCA[31]. In SCA, datasets are required to reflect characteristics of competitive sports, which is a guarantee for training deep learning models. Generally, competitive sports content is a series of diversified, high professional and ulti- mate actions. Unfortunately, existing trending human motion Shenglan Liu, Xiang Liu, Lin Feng, Lianyu Hu, Dong Jiang, Aibin Zhang, Yang Liu are with Faculty of Electronic Information and Elec- trical Engineering, Dalian University of Technology, Dalian, Liaoning, 116024 China. Email: ({liusl, hly19961121, s201461147, 18247666912, dlut liuyang}@mail.dlut.edu.cn, [email protected], [email protected]). Gao Huang is with Faculty of Tsinghua University of Technology, Beijing, China. Email: [email protected]. H. Qiao is with the State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190 China. E-mail: [email protected]. This work is supported in part by The National Key Research and Develop- ment Program of China (2017YFB1300200) National Natural Science Foun- dation of People’s Republic of China (No. 61672130, 61602082, 91648205, 31871106), the National Key Scientific Instrument and Equipment Develop- ment Project (No. 61627808), the Development of Science and Technology of Guangdong Province Special Fund Project Grants (No. 2016B090910001), the LiaoNing Revitalization Talents Program (No.XLYC1086006). datasets (e.g. HMDB51[18], UCF50[34]) or action datasets of human sports (e.g. MIT Olympic sports[28], Nevada Olympic sports[24]) are not quite representative of the richness and complexity of competitive sports. The limitations of current datasets could be summarised as follows: (1) Current tasks of human action analysis are insufficient, which is limited to action types and contents of videos. (2) In most classification tasks, discriminant of an action largely depends on both static human pose and background; (3) Video segmentation, as an important task of video understanding (including classification and assessment in SCA ), is rarely discussed in current datasets. To address the above issues, this paper proposes a figure skating dataset called FSD-10. FSD-10 consists of 1484 figure skating videos with 10 different actions manually labeled. These skating videos are segmented from around 80 hours of competitions of worldwide figure skating championships 1 in 2017-2018. FSD-10 videos range from 3 seconds to 30 seconds, and the camera is moving to focus the skater to ensure that he (she) appears in each frame during the process of actions. Compared with existing datasets, our proposed dataset has several appealing properties. First, actions in FSD-10 are complex in content and fast in speed. For instance, the complex 2-loop-Axel jump in Fig. 1 is finished in only about 2s. It’s worth note that the jump type heavily depends on the take off process, which is a hard-captured moment. Second, actions of FSD-10 are original from figure skating competitions, which are consistent in type and sports environment (only skater and ice around). The above two aspects create difficulties for machine learning model to conclude the action types by a single pose or background. Third, FSD-10 can be applied to standardized action quality assessment (AQA) tasks. The AQA of FSD-10 actions depends on base value (BV) and grade of execution (GOE), which increases the difficulty of AQA by both action qualities and rules of International Skating Union (ISU) 2 . Last, FSD-10 can be extended to more tasks than the mentioned datasets. Along with the introduction of FSD-10, we introduce a key frame indicator called human pose scatter (HPS). Based on HPS, we adopt key frame sampling to improve current video classification methods and evaluate these methods on FSD- 1 ISU Grand Prix of Figure Skating, European Figure Skating Championships, Four Continents Figure Skating Championships (https://www.isu.org/figure-skating) 2 One action appearing in second half of program will earn extra rate of 10% score of BV than that appearing in first half of program. This means that same action may get different BV score because of the rules. arXiv:2002.03312v1 [cs.CV] 9 Feb 2020

Upload: others

Post on 29-Sep-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 1

FSD-10: A Dataset for Competitive Sports ContentAnalysis

Shenglan Liu Member, IEEE, Xiang Liu, Gao Huang, Lin Feng, Lianyu Hu, Dong Jiang, Aibin Zhang, Yang Liu,Hong Qiao Fellow, IEEE

Abstract—Action recognition is an important and challeng-ing problem in video analysis. Although the past decade haswitnessed progress in action recognition with the developmentof deep learning, such process has been slow in competitivesports content analysis. To promote the research on actionrecognition from competitive sports video clips, we introduce aFigure Skating Dataset (FSD-10) for finegrained sports contentanalysis. To this end, we collect 1484 clips from the worldwidefigure skating championships in 2017-2018, which consist of 10different actions in men/ladies programs. Each clip is at a rateof 30 frames per second with resolution 1080 × 720. These clipsare then annotated by experts in type, grade of execution, skaterinfo, .etc. To build a baseline for action recognition in figureskating, we evaluate state-of-the-art action recognition methodson FSD-10. Motivated by the idea that domain knowledge isof great concern in sports field, we propose a keyframe basedtemporal segment network (KTSN) for classification and achieveremarkable performance. Experimental results demonstrate thatFSD-10 is an ideal dataset for benchmarking action recognitionalgorithms, as it requires to accurately extract action motionsrather than action poses. We hope FSD-10, which is designedto have a large collection of finegrained actions, can serve asa new challenge to develop more robust and advanced actionrecognition models.

I. INTRODUCTION

Due to the popularity of media-sharing platforms(e.g.YouTube), sports content analysis (SCA[31]) has becomean important research topic in computer vision [39][10][30].A vast amount of sports videos are piled up in computerstorage, which are potential resources for deep learning. Inrecent years, many enterprises (e.g. Bloomberg, SAP) havefocus on SCA[31]. In SCA, datasets are required to reflectcharacteristics of competitive sports, which is a guarantee fortraining deep learning models. Generally, competitive sportscontent is a series of diversified, high professional and ulti-mate actions. Unfortunately, existing trending human motion

Shenglan Liu, Xiang Liu, Lin Feng, Lianyu Hu, Dong Jiang, AibinZhang, Yang Liu are with Faculty of Electronic Information and Elec-trical Engineering, Dalian University of Technology, Dalian, Liaoning,116024 China. Email: ({liusl, hly19961121, s201461147, 18247666912,dlut liuyang}@mail.dlut.edu.cn, [email protected], [email protected]).

Gao Huang is with Faculty of Tsinghua University of Technology, Beijing,China. Email: [email protected].

H. Qiao is with the State Key Laboratory of Management and Control forComplex Systems, Institute of Automation, Chinese Academy of Sciences,Beijing 100190 China. E-mail: [email protected].

This work is supported in part by The National Key Research and Develop-ment Program of China (2017YFB1300200) National Natural Science Foun-dation of People’s Republic of China (No. 61672130, 61602082, 91648205,31871106), the National Key Scientific Instrument and Equipment Develop-ment Project (No. 61627808), the Development of Science and Technologyof Guangdong Province Special Fund Project Grants (No. 2016B090910001),the LiaoNing Revitalization Talents Program (No.XLYC1086006).

datasets (e.g. HMDB51[18], UCF50[34]) or action datasets ofhuman sports (e.g. MIT Olympic sports[28], Nevada Olympicsports[24]) are not quite representative of the richness andcomplexity of competitive sports. The limitations of currentdatasets could be summarised as follows: (1) Current tasksof human action analysis are insufficient, which is limited toaction types and contents of videos. (2) In most classificationtasks, discriminant of an action largely depends on both statichuman pose and background; (3) Video segmentation, as animportant task of video understanding (including classificationand assessment in SCA ), is rarely discussed in currentdatasets.

To address the above issues, this paper proposes a figureskating dataset called FSD-10. FSD-10 consists of 1484 figureskating videos with 10 different actions manually labeled.These skating videos are segmented from around 80 hoursof competitions of worldwide figure skating championships1 in 2017-2018. FSD-10 videos range from 3 seconds to 30seconds, and the camera is moving to focus the skater to ensurethat he (she) appears in each frame during the process ofactions. Compared with existing datasets, our proposed datasethas several appealing properties. First, actions in FSD-10 arecomplex in content and fast in speed. For instance, the complex2-loop-Axel jump in Fig. 1 is finished in only about 2s. It’sworth note that the jump type heavily depends on the take offprocess, which is a hard-captured moment. Second, actions ofFSD-10 are original from figure skating competitions, whichare consistent in type and sports environment (only skaterand ice around). The above two aspects create difficulties formachine learning model to conclude the action types by asingle pose or background. Third, FSD-10 can be applied tostandardized action quality assessment (AQA) tasks. The AQAof FSD-10 actions depends on base value (BV) and grade ofexecution (GOE), which increases the difficulty of AQA byboth action qualities and rules of International Skating Union(ISU)2. Last, FSD-10 can be extended to more tasks than thementioned datasets.

Along with the introduction of FSD-10, we introduce a keyframe indicator called human pose scatter (HPS). Based onHPS, we adopt key frame sampling to improve current videoclassification methods and evaluate these methods on FSD-

1ISU Grand Prix of Figure Skating, European Figure SkatingChampionships, Four Continents Figure Skating Championships(https://www.isu.org/figure-skating)

2One action appearing in second half of program will earn extra rate of10% score of BV than that appearing in first half of program. This meansthat same action may get different BV score because of the rules.

arX

iv:2

002.

0331

2v1

[cs

.CV

] 9

Feb

202

0

Page 2: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 2

Warm up𝜁𝐻𝑃𝑆(82.89)

Take off𝜁𝐻𝑃𝑆(187.15)

Spin in the air 𝜁𝐻𝑃𝑆 (61.88)

Landing𝜁𝐻𝑃𝑆(152.63)

Fig. 1. An over view of the figure skating dataset. It is illustrated that warm up, take off, spinning in the air and landing are 4 steps of an Axel jump. ζHPS

indicates the human pose scatter of a gesture, which is introduced in Sec IV-B.

TABLE IA SUMMARY OF EXISTING ACTION DATASETS

Dataset Year Actions Clips Classification Score Temporal Segmentation

UCF Sports[29] 2009 9 14-35√

- -Hollywood2[21] 2009 12 61-278

√- -

UCF50[34] 2010 50 min. 100√

- -HMDB51[18] 2011 51 min. 101

√- -

MIT Olympic sports[28] 2015 2 150-159√ √

-Nevada Olympic sports[25] 2017 3 150-370

√ √-

AQA-7[24] 2018 7 83-370√ √

-FSD-10 2019 10 91-233

√ √ √

10. Furthermore, experimental results validate that key framesampling is an important approach to improve performance offrame-based model in FSD-10, which is in concert with cogni-tion rules of human in figure skating. The main contributionsof this paper can be summarised as follows.

• To our best knowledge, FSD-10 is the first challengingdataset with high-speed, motion-focused and complexsport actions in competitive sports field, which introducesmultiple tasks for computer vision, such as (fine grained)classification, long/short temporal human motion segmen-tation, action quality assessments, and program contentassessment in time to music.

• To set a baseline for future achievements, we also bench-mark state-of-the-art sport classification methods on FSD-10. Besides, the key frame sampling is proposed tocapture the pivotal action details in competitive sports,which achieves better performance than state-of-the-artmethods in FSD-10.

In addition,compared to current datasets, we hope FSD-10will be a standard dataset which promotes the research onbackground-independent action recognition, allowing modelsto be more robust to unexpected events and generalize to novelsituations and tasks.

II. RELATED WORKS

Current action datasets could be divided into human actiondataset (HAD) and human professional sports dataset (HPSD),which are listed in Tab. I. HAD consists of different kinds ofbroadly defined human motion clips, such as haircut and longjump. For example, UCF101[34] and HMDB[18] are typicalcases of HAD, which have greatly promoted the process ofvideo classification techniques. In contrast, HPSD is a series ofcompetitive sports actions, which is proposed for action qualityassessment (AQA) tasks except classification. MIT Olympicsports[28] and Nevada Olympic sports[25] are examples ofHPSD, which are derived from Olympic competitions.

Classification in HPSD is important to attract people’s atten-tion and to highlight athlete’s performance, and even to assistreferees [16][46][1]. The significant difference between videoclassification and image classification is logical connectionrelationship in video clips [2], which involve motion besidesform (content) of frames. Therefore, human action classifi-cation task should be implemented based on temporal humaninformation without background rather than static action frameinformation. H.kuehne et al. [18] mentioned that the recogni-tion rate of static joint locations alone in UCF sports [29]dataset is above 98% while the use of joint kinematics is notnecessary. The result above obviously violates to the principle

Page 3: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 3

ChComboSpin4

FlyCamelSpin4

ChoreoSequence1

StepSequence3

3Axel

3Flip

3Loop

3Lutz

2Axel 3Lutz_3Toeloop

Fig. 2. 10 figure skating actions included in FSD-10 shown with 5 frames. The color of frame borders specifies to which action category they belong (Spin,Sequence, Jump) and the labels specifies their action types.

of video classification task. Thus, both motion and pose areindispensable modalities in human action classification [7].

Besides classification of HPSD, AQA is an important butunique task in HPSD. Some AQA datasets have been proposed,such as MIT Olympic sports dataset[28], Nevada Olympicsports dataset[25] and AQA-7[24], which predict the actionperformance of diving, gymnastic vault, etc. In this field,Parmar et al. [26] adopts C3D features combining regressionwith the help of SVR[6] (LSTM[12]), which gets satisfactoryresults. The development direction of HPSD is clearly givenby the above works while AQA still remains to be devel-oped as the materials of these datasets is still inadequate.It could be concluded that the functions of datasets aregradually expanded with the development of video techniques[17][9][27][43][32]. However, it lacks temporal segmentationrelated datasets in recent years. Our proposed FSD-10 is aresponse to these issues. As far as we know, FSD-10 is thefirst dataset which gather actions task types including actionclassification, AQA and temporal segmentation.

III. FIGURE SKATING DATASET

In this section, we describe details regarding the setup andprotocol followed to capture our dataset. Then, we discuss thetemporal segmentation and assessment tasks of FSD-10 andits future extensions.

A. Video Capture

In FSD-10, all the videos are captured from the WorldFigure Skating Championships in 2017-2018, which originatesfrom high-definition record on YouTube. The source videomaterials have large ranges of fps (frame per second) andimage size, which are standardized to 30 fps and 1080 × 720to ensure the consistency of the dataset respectively. As forthe content, the programs of figure skating can be divided intofour types, including men short program, ladies short program,men free skating and ladies free skating. During the skating,camera is moving with the skater to ensure that the currentaction can be located center in videos. Besides, auditoriumand ice ground are collectively referred to the background,which are similar and consistent in all videos.

B. Segmentation and Annotation

To meet the demand of deep learning and other machinelearning models, video materials are manually separated intoaction clips by undergraduate students. Each clip is rangingfrom 3s to 30s, the principe of the segmentation is to com-pletely reserve action content of videos, in other words, warm-up and ending actions are also included. Actions type, basevalue (BV), grade of execution (GOE), skater name, skatergender, skater age, coach and music are recorded in eachclip. A detailed record example of one clip is shown in Tab.II. Unfortunately, the distribution of actions is out of tune,

Page 4: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 4

Fig. 3. Number of clips per action class. The horizontal axis indicates action types while the vertical axis indicates action amout.

TABLE IIA SINGLE CLIP EXAMPLE OF FSD-10

Action Value

Type 2AxelBV 3.3/3.63

GOE 0.58Skater name Starr ANDREWS

Skater gender FSkater age 16

Coach Derrick Delmore, Peter KongkasemMusic Fever performed by Beyonce

as different kinds of actions are totally 197 and the numberof each class of action instances is from 1 to 223. To avoidinsufficient training, our FSD-10 is composed of the top 10actions (as shown in Fig. 2, Fig. 3). Generally speaking, theseselected actions could fall into three categories: jump, spinand sequence. And in details, ChComboSpin4, FlyCamelSpin4,ChoreoSequence1, StepSequence3, 2Axel, 3Loop, 3Flip, 3Axel,3Lutz and 3Lutz 3Toeloop are included.

C. Action Discriminant Analysis

Compared with existing sports, figure skating is hardlydistinguished in action types. The actions in figure skatinglook like tweedledum and tweedledee, which are difficult todistinguish in classification task. (1) It is difficult to judge theaction type by human pose in a certain frame or a few frames.For instance, Loop jump and Lutz jump are similar in warm-up period, it is impossible to distinguish them beyond take-offprocess. (2) The difference between some actions exists on asubtle pose in fixed frames. For example, Flip jump and Lutzjump are both picking-ice actions, which are same in take-offwith left leg and picking ice with right toe. The only subtledifference of the two types of jump is that takes-off process ofFlip jump is by inner edge while Lutz jump is by outer edge.(3) The spin direction is uncertain. The rotational direction ofmost skaters is anticlockwise while that of others is opposite(e.g. Denise Biellmann). These factors of the above issuesmake classification task of FSD-10 a challengeable problem.

D. Temporal Segmentation and Assessment

• Action Quality Assessment (AQA). AQA would beemerged as an imperative and challengeable issue infigure skating, which is used to evaluate performancelevel of skaters in skating programs by BV score plusGOE score. As shown in Sec. III-B, BV and GOE areincluded in our dataset, which are mainly depend onaction types and skater’s action performance respectively.The key points of AQA are summarised as follows. (1)BV is depend on action types and degree of actiondifficulty. Besides, 10% bonus BV score is appendedon the second half of a program. (2) During the aerialprocess, insufficient number of turns and raising handscause GOE deduction and bonus respectively. (3) It iscaused GOE deduction of one action by hand support,turn over, paralleling feet and trip during the landingprocess.

• Action Temporal Segmentation(ATS). It is indicatedthat ATS is also a significant issue of FSD-10. A standardaction is of certain stages. For example during the processof a jump action, four stages are commonly included,which are preparation, take-off process, spinning in theair and landing process. Separating these steps is a keytechnical issue for AQA and further related works.

• Long Sequence Segmentation and Assessment (LSSA).Another important issue is LSSA, which is differentfrom AAS in sensitivity of actions order. Action orderand relation need to be considered in LSSA, which issummarised as follows. (1) LSSA is the total of all thesingle action score, which is calculated similar to AQAfor each action. (2) It is caused a fail case 3 when thetwice appearance of same action.

E. Future Extension

Our proposed FSD-10 is sensitive to domain knowledge andabundant in practical tasks. Besides the tasks in Sec. III-D ,

3The twice appearance action will have a zero score whether it is successfulor not.

Page 5: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 5

FSD-10 is convenient to be extended into other fields. Forexample, pair skating is also a worth noting program in figureskating. In the pair skating, to throw one skater by another isa classical item, and the action in the air could be illustratedas a corresponding jump action in men/ladies programs. Totransfer single skating models to pair skating ones is aninteresting open problem, which could be deemed as a kindof transfer learning[23][45] or meta learning[19][14]. Besides,action reasoning[8][38] is an interesting issue. For example,it is straightforward to conclude a 3Lutz-3Toeloop jump ifsingle 3Lutz jump and 3toeloop jump have been recognized. Inaddition, coordination between the action and the other sportselements such as program content assessment by rhythmicalactions in time to music is also a focused issue. Fortunately,the corresponding music of actions is included in FSD-10. Itis to be explored whether the music ups and downs is relatedto the action rhythm, which is an exciting cross-data-modalitylearning task.

IV. KEYFRAME BASED TEMPORAL SEGMENT NETWORK(KTSN)

In this section, we give a detailed description of ourkeyframe based temporal segment network. Specifically, wefirst discuss the motivation of key frame sampling in Sec.IV-A. Then, sampling method of key frame is proposed inSec. IV-B and IV-C. Finally, network structure of KTSN isdetailedly introduced in Sec. IV-D.

A. Motivation of Key Frame Sampling

To solve the problem of long-range temporal modeling, asampling method called segment based sampling is proposedby temporal segment networks(TSN)[40]. The main idea ofsegment based sampling is “the dense recoded videos areunnecessary for content which changes slowly”. The methodis essentially a kind of sparse and global sampling, whichis significant in improving recognition performance of long-range videos. But an intractable problem is caused that it isunbefitting for fast changing videos. Especially for tasks likefigure skating, which is complex in content and fast in speed.It is illustrated in Sec. III-C that actions of figure skating aresimilar in most of frames, and the only difference lies oncertain frames. Therefore, for sampling problem, we proposea method to sample key frames by domain knowledge and tosample other frames by segment based sampling. It is thenfocus on how to sample key frames in segmentation basedsampling.

The problem motivates us to explore an indicator of sam-pling by domain knowledge in figure skating. For example,by observing the process of jump, we find that arms and legsmove following specific rules during the process. Thus, we areencouraged to define an indicator using anatomical keypointsto measure this kind of rules, which is called Human PoseScatter and is illustrated in the next section.

B. Human Pose Scatter (HPS)

Before key frame sampling, we will introduce HPS for apreparation. HPS is an indicator which is denoted by the

scatter of arms and legs relative to the body. The mainpositions of human (left eye, right eye, etc. ) are marked bythe open-pose algorithm, which is shown in Fig. 4. Next, HPSis introduced in the rest of this section.

ҧ𝑥

𝑈

𝑥1

Fig. 4. The sketch map of human pose scatter. A 18-anatomical-keypointsrecognition algorithm called open-pose[4][3][33][41] is adopted to get thespatial position and anatomical structure of human in one frame.

As shown in Fig. 4, given a frame matrix of the respondinganatomical structure, X = [x1, · · · , xn] ∈ RD×n (in thispaper D = 2 and n = 18), in which each row corresponds to aparticular feature while each column corresponds to a humananatomical keypoint, and x is the column mean value of X .Then, X can be defined as X = [x1− x, x2− x, · · · , xi− x].Motivated by the principal component analysis (PCA[42]),HPS can be defined by Eq. 1. Projection matrix U (UD×d =[u1, u2, · · · , ud], satisfying UTU = I , which leads I − UUTthe idemfactor) can be calculated by eigenvalue decomposition(EVD) of XXT .

ζHPS =

n∑i=1

∥∥(xi − x)− UUT (xi − x)∥∥22

(1)

Actually, Eq. 1, only a formal expression of ζHPS , canbe easily rewritten as Eq. 2 to simplify calculation. In Eq.2, Tr

[XXT

]and Tr

[UT XXTU

]=∑dj=1 λj

4 have beencalculated during the EVD process. Therefore, Eq. 2 offers anefficient approach for calculation of ζHPS .

ζHPS =

n∑i=1

∥∥(I − UUT )(xi − x)∥∥22

=

n∑i=1

(xi − x)T (I − UUT )(xi − x)

=Tr(XXT )−d∑j=1

λj

(2)

ζHPS is the sum of variance towards the principal axisof human anatomical keypoints, which is used to reflectthe degree of human pose stretching. Therefore, ζHPS is

4λj is the j-th eigenvalue of covariance matrix XXT , in our paper d = 1.

Page 6: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 6

a tremendous help in finding key frames of sequences. Forexample, during a process of jump, ζHPS of spin in the airprocess is lower than it of the take off process, and extremepoint maps a turning point responding to some certain actions.This can be referred to Fig. 1.

C. Key Frame Sampling

𝜁𝐻𝑃𝑆

𝑇

“The specific key frame”

O𝑗∗

𝐺𝑗∗,𝑗∗−𝛿

𝐺𝑗∗,p𝐺𝑗∗,𝑗∗+𝛿

𝜁𝐻𝑃𝑆

t

Fig. 5. The sketch map of key frame sampling. A 2-Axel jump is performedby Alina Ilnazovna Zagitova, which is a same action with Fig. 1. t representsthe time variable.

Key frame sampling is an important issue in video analysisof figure skating, and the key frames should contain mostdiscriminant information of one video. In the figure skatingtask, the fast changing motion frames are distinctly importantfor jump actions. As illustrated in Sec. IV-B, the kind offast changing motion could be quantified as ζHPS by usinganatomical keypoints. Therefore, in a T frames clip of avideo, the frame corresponding to extreme frame Oj∗ , whichsatisfies j∗ = argmax

jζjHPS , is defined as “the specific key

frame”, where δ indicates the near frame radius of Oj∗ , j ∈{δ + 1, · · · , T − δ}. In addition, as nearby δ frames of Oj∗ ,Gj∗ = {Gj∗,j∗−δ, · · · , Gj∗,p, · · · , Gj∗,j∗+δ} are defined as“key frames”, where p ∈ {j∗ − δ, · · · , j∗ + δ} \j∗(see Fig. 5).In a multi-action clip, we will get Olj∗ , where l = 1, 2, · · · , L,L indicates the number of extreme frames. Then, key framesampling could be described as follows. First, finding thespecific key frame(s) in action sequences. Second, samplingthe video from the specific key frame(s) and its nearby framesusing the similar approach of TSN. In one batch, TSN samplesa random frame while key frame sampling involves a randomframe Glj∗,p in Gj∗ ∪Oj∗ .

D. Framework Structure

An effective and efficient video-level framework, coinedkeyframe based Temporal Segment Network (KTSN, Fig. 6),is designed with the help of key frame sampling. Comparedwith single frame or short frame pieces, KTSN works on

short snippets sampled from the videos. To make these shortsnippets representative for the entire videos as well as to keepcomputational consumption steerable, the video is dividedinto several segments first. Then, sampling is continued oneach piece of snippets. As illustrated in IV-A, our samplingmethods are divided to two parts, which are segment basedsampling[40] and key frame sampling respectively. With thesampling method, each piece contributes its snippet-level pre-dictions of actions classes. And these snippet-level predictionsare aggregated as video-level scores by a consensus function.The video-level prediction is more accurate than any singlesnippet-level predictions for the reason that it captures detailsof the whole videos. During the process of training, the modelis iteratively updated in video-level as follows.

We divide a video into k segments (Vvideo ={α1, α2, · · · , αk} of equal duration pieces. One frame Tkis randomly selected from its corresponding segment αk.Similarly, One snippet Glj∗,p is randomly selected from itscorresponding Glj∗∪Olj∗ , which have been proposed in sectionIV-B and IV-C. The sequence of snippets (T1, T2, · · · , Tk andG1j∗,p, G

2j∗,p, · · · , GLj∗,p) makes up of the model of keyframe

based temporal segment network as follows.

KTSN(T1,T2, · · · , Tk, G1j∗,p, G

2j∗,p, · · · , GLj∗,p) =

P(G(F(T1;W ),F(T2;W ), · · · ,F(Tk;W )),

G(F(G1j∗,p;W ),F(G2

j∗,p;W ), · · · ,F(GLj∗,p;W )))(3)

In Eq. 3, F(Tk;W ) (or F(Glj∗,p;W )) is the function onbehalf of a ConvNet[20] with model parameters W which playsan important role in the snippet Tk (Glj∗,p) and obtain a con-sensus of classification hypothesis. It is given the prediction ofthe whole video by an aggregation function G, which combinesall the output of individual snippets. Based on the value ofG, P works out the classification results. In this paper, P isrepresented as softmax function[22].

V. EXPERIMENTS

In order to provide a benchmark for our FSD-10 dataset, weevaluate various approaches under three different modalities:RGB, optical flow and anatomical keypoints (skeleton). Wealso conduct experiments on cross dataset validation. Thefollowing describes the details of our experiments and results.

A. Evaluations of Methods on FSD-10The state-of-the-art classification methods of human action

are evaluated on FSD-10 in this section. In order to makethe experiments representative in action recognition field, 2Dmethods and 3D methods are selected respectively. TSN (orKTSN) is selected as a 2D method, which is a typical twostream structure. As for 3D methods, two-stream inflated3D ConvNet (I3D) with Inception-v1[36], Inception-v3[37]and Resnet-50, Resnet-101[11] are involved for comparison.Besides, anatomical keypoints related methods (st-gcn[44],densenet[15]) are tested on our dataset for reference. As listedin Tab. III, our proposed KTSN achieves state-of-the-art resultson FSD-10. KTSN method is raised as a baseline for this high-speed dataset.

Page 7: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 7

Key Frame Sampling

Segment based Sampling

Key-frame based Temporal Segment network

𝑇1

𝑇2

𝑇3

𝐺𝑗∗,𝑝1

ℱ(𝑇1;𝑊)

ℱ(𝑇2;𝑊)

ℱ(𝑇3;𝑊)

ℱ(𝐺𝑗∗,𝑝1 ;𝑊)

𝒢 ℱ(𝑇1;𝑊 , ℱ(𝑇2;𝑊),ℱ(𝑇3;𝑊))

𝒢(ℱ(𝐺𝑗∗,𝑝1 ;𝑊))

𝒫 𝒢 ℱ(𝑇1;𝑊 , ℱ(𝑇2;𝑊 ,

ℱ(𝑇3;𝑊)), 𝒢(ℱ(𝐺𝑗∗,𝑝1 ;𝑊)))

Red layer

Green layer

Blue layer

Video Sequence

Optical flow x

Optical flow y

Pooling

Other

Softmax

Convolution

Input & Output

𝒢( ), 𝒫( )

Fig. 6. Keyframe based temporal segment network. One input video is divided into k segments (here we show the k = 3 case) and a short snippet israndomly selected from each segment. Meanwhile, n key frames are selected from the Video (here L = 1). The snippets are represented by modalities such asRGB frames and Optical flow. Then these snippets are fused by an aggregation function G. Last, P combines the outputs and makes video-level predictions.ConvNets on all snippets share parameters.

TABLE IIIPERFORMANCE OF BASELINE CLASSIFICATION MODELS EVALUATES WITH

FSD-10. THE SHORT NAMES OF THESE METHODS ARE TWO-STREAMINFLATED 3D CONVNET[5] (INCEPTION-V1[36], INCEPTION-V3[37],RESNET-50[11] AND RESNET-101[11]), SPATIAL TEMPORAL GRAPH

CONVOLUTION NETWORKS (ST-GCN[44]), DENSENET WITH ANATOMICALKEYPOINTS NETWORKS[15], TEMPORAL SEGMENT NETWORK[40] AND

KEYFRAME BASED TEMPORAL SEGMENT NETWORK RESPECTIVELY.DENSENET+ DENOTES DENSENET WITH CENTRALIZATION. BESIDES,

TSN(7) REPRESENTS TSN(k=7) AND KTSN(6,1) REPRESENTSKTSN(k=6, L=1) CORRESPONDING TO SEC. IV-D, SIMILARLY

HEREINAFTER.

RGB Optical flow Skeleton Fusion

Inception-v1[36] 43.765 52.706 - 55.052Inception-v3[37] 50.118 57.245 - 59.624

Resnet-50[11] 62.550 58.824 - 64.941Resnet-101[11] 78.823 - 61.549 79.851

St-gcn[44] - - 72.429 -DenseNet[15] - - 74.353 -

DenseNet+[15] - - 79.765 -

TSN(7)[40] 53.176 75.165 - 76.000KTSN(6, 1) 56.941 76.941 - 77.412

TSN(9)[40] 59.294 80.235 - 80.235KTSN(8, 1) 59.529 80.471 - 80.471

TSN(11)[40] 59.294 82.118 - 82.118KTSN(10, 1) 63.294 82.588 - 82.588

B. Evaluations between TSN and KTSN

The experimental results in Tab. III show that key framesampling is effective in sparse sampling. As illustrated above,key frame is crucial for competitive sports classification tasks.Especially for jump actions, it is vital to choose take offframes, which enhances discriminant of different jump actions.An example of key frame sampling of jump actions is shownin Fig. 5. By catching these key frames instead of randomsampling, KTSN achieves better classification performance

than TSN[40].

C. Cross-dataset Validations

To verify motion characteristic of a sport video, the cross-dataset validation is conducted on both classical datasetUCF101 and FSD-10. RGB and optical flow features are twotypical ways for action recognition. Generally speaking, theRGB feature based method places emphasis on human posewhile the optical flow based method attaches importance totransformation between actions. It is illustrated in Tab. IVthat the performance differences between RGB and Opticalflow of FSD-10 is greater than UCF101, which indicatesmotion is more important than human static pose in FSD-10. Besides, in sampling stage, FSD-10 is more sensitive tofine-grained segmentation, as the performance raises 5.4%compared with 2.5% in UCF101 (3-segment-sampling to 9-segment-sampling). In FSD-10, actions are sensitive to itsmotion rather than its pose, which is consistent with theoriginal intention of the dataset.

TABLE IVRESULTS OF CROSS-DATASET GENERALIZATION. UCF101 AND FSD-10

ARE TRAINED WITH TSN(k=3) AND TSN(k=9) RESPECTIVELY.

Dataset Methods RGB Optical flow Fusion

UCF101 TSN(3) 85.673 85.170 92.466TSN(9) 86.217 89.713 94.909

FSD-10 TSN(3) 48.941 75.765 76.765TSN(9) 59.294 82.118 82.118

VI. CONCLUSION

In this paper, we build an action dataset for competitivesports analysis, which is characterised by high action speed

Page 8: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 8

and complex action content. We find that motion is morevaluable than form (content and background) in this task.Therefore, compared with other related datasets, our datasetfocuses on the action itself rather background. Our datasetcreates many interesting tasks, such as fine-grained actionclassification, action quality assessment and action temporalsegmentation. Moreover, the dataset could even be extendedto pair skating in transfer learning and coordination betweenaction and music in cross-data-modality learning.

We also present a baseline for action classification in FSD-10. Note that domain knowledge is significant for competitiveaction classification task. Motivated by this idea, we adopt key-frame based temporal segmentation sampling. It is illustratedthat action sampling should focus more on related frames ofmotion. The experimental results demonstrate that the keyframe is effective for skating tasks.

In the future, more diversified action types and tasks areintended to be included to our dataset, and more interestingaction recognition methods are likely to be proposed with thehelp of FSD-10. We believe that our dataset will promote thedevelopment of action analysis and related research topics.

SUPPLEMENTARY MATERIAL

1. Training of KTSN

Architectures Inception network, chosen in KTSN, is amilestone in CNN classifier. The main characteristics of in-ception network is the inception module, which is designed toprocess the features in different dimensionality (such as the1×1, 3×3 and 5×5 convolution core) parallelly. The result ofInception network is a better representation than single-layer-networks, which is excellent in local topological structure.Specifically, Inception v3 is selected in the experiments asfollows.

Input Our model is based on spatial and temporal level,which are corresponding to RGB frames[35] and opticalflow[13] respectively. RGB mode is an industrial standardof color, combining Red, Green and Blue into chromaticframes, which is also called the spatial flow. Besides, opticalflow mode derives from calculating the differences betweenadjacent frames, which indicates temporal flow.

2. Confusion Matrix of Actions

It’s significant to involve domain knowledge in competitivesports content analysis. It is illustrated the confusion relation-ships between actions by examples in Sec. 3.3, and this part isa validation, as shown in Fig. 7. First, the skating spin obtainsthe best classification results, this is because it could judgea skating spin of video precisely with a few isolated framesusing deep networks. Second, the skating jump is the mostdifficult to distinguish. It’s worth note that up to 24% Lutzjump actions are identified as Flip jump actions.

REFERENCES

[1] Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Paul Natsev, GeorgeToderici, Balakrishnan Varadarajan, and Sudheendra Vijayanarasimhan.Youtube-8m: A large-scale video classification benchmark. arXivpreprint arXiv:1609.08675, 2016.

Fig. 7. The confusion matrix results of FSD-10

[2] Alan C Bovik. Handbook of image and video processing. Academicpress, 2010.

[3] Zhe Cao, Gines Hidalgo, Tomas Simon, Shih-En Wei, and Yaser Sheikh.OpenPose: realtime multi-person 2D pose estimation using Part AffinityFields. In arXiv preprint arXiv:1812.08008, 2018.

[4] Zhe Cao, Tomas Simon, Shih-En Wei, and Yaser Sheikh. Realtimemulti-person 2d pose estimation using part affinity fields. In CVPR,2017.

[5] Joao Carreira and Andrew Zisserman. Quo vadis, action recognition?a new model and the kinetics dataset. In proceedings of the IEEEConference on Computer Vision and Pattern Recognition, pages 6299–6308, 2017.

[6] Harris Drucker, Christopher JC Burges, Linda Kaufman, Alex J Smola,and Vladimir Vapnik. Support vector regression machines. In Advancesin neural information processing systems, pages 155–161, 1997.

[7] Hany El-Ghaish, Mohamed E Hussein, Amin Shoukry, and Rikio Onai.Human action recognition based on integrating body pose, part shape,and motion. IEEE Access, 6:49040–49055, 2018.

[8] Matthew L Ginsberg and David E Smith. Reasoning about action i: Apossible worlds approach. Artificial intelligence, 35(2):165–195, 1988.

[9] Chunhui Gu, Chen Sun, David A Ross, Carl Vondrick, CarolinePantofaru, Yeqing Li, Sudheendra Vijayanarasimhan, George Toderici,Susanna Ricco, Rahul Sukthankar, et al. Ava: A video dataset of spatio-temporally localized atomic visual actions. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition, pages 6047–6056, 2018.

[10] Joachim Gudmundsson and Michael Horton. Spatio-temporal analysisof team sports. ACM Computing Surveys (CSUR), 50(2):22, 2017.

[11] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deepresidual learning for image recognition. In Proceedings of the IEEEconference on computer vision and pattern recognition, pages 770–778,2016.

[12] Sepp Hochreiter and Jurgen Schmidhuber. Long short-term memory.Neural computation, 9(8):1735–1780, 1997.

[13] Berthold KP Horn and Brian G Schunck. Determining optical flow.Artificial intelligence, 17(1-3):185–203, 1981.

[14] Kyle Hsu, Sergey Levine, and Chelsea Finn. Unsupervised learning viameta-learning. arXiv preprint arXiv:1810.02334, 2018.

[15] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian QWeinberger. Densely connected convolutional networks. In Proceedingsof the IEEE conference on computer vision and pattern recognition,pages 4700–4708, 2017.

[16] Andrej Karpathy, George Toderici, Sanketh Shetty, Thomas Leung,Rahul Sukthankar, and Li Fei-Fei. Large-scale video classification withconvolutional neural networks. In Proceedings of the IEEE conferenceon Computer Vision and Pattern Recognition, pages 1725–1732, 2014.

[17] Will Kay, Joao Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier,Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back,Paul Natsev, et al. The kinetics human action video dataset. arXivpreprint arXiv:1705.06950, 2017.

Page 9: JOURNAL OF LA FSD-10: A Dataset for Competitive Sports Content Analysis · 2020. 2. 11. · sports content analysis. To promote the research on action recognition from competitive

JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 9

[18] Hildegard Kuehne, Hueihan Jhuang, Estıbaliz Garrote, Tomaso Poggio,and Thomas Serre. Hmdb: a large video database for human motionrecognition. In 2011 International Conference on Computer Vision,pages 2556–2563. IEEE, 2011.

[19] Kwonjoon Lee, Subhransu Maji, Avinash Ravichandran, and StefanoSoatto. Meta-learning with differentiable convex optimization. InProceedings of the IEEE Conference on Computer Vision and PatternRecognition, pages 10657–10665, 2019.

[20] Shih-Chung B Lo, Heang-Ping Chan, Jyh-Shyan Lin, Huai Li,Matthew T Freedman, and Seong K Mun. Artificial convolution neuralnetwork for medical image pattern recognition. Neural networks, 8(7-8):1201–1214, 1995.

[21] Marcin Marszałek, Ivan Laptev, and Cordelia Schmid. Actions incontext. In CVPR 2009-IEEE Conference on Computer Vision & PatternRecognition, pages 2929–2936. IEEE Computer Society, 2009.

[22] Roland Memisevic, Christopher Zach, Marc Pollefeys, and Geoffrey EHinton. Gated softmax classification. In Advances in neural informationprocessing systems, pages 1603–1611, 2010.

[23] Maxime Oquab, Leon Bottou, Ivan Laptev, and Josef Sivic. Learning andtransferring mid-level image representations using convolutional neuralnetworks. In Proceedings of the IEEE conference on computer visionand pattern recognition, pages 1717–1724, 2014.

[24] Paritosh Parmar and Brendan Morris. Action quality assessment acrossmultiple actions. In 2019 IEEE Winter Conference on Applications ofComputer Vision (WACV), pages 1468–1476. IEEE, 2019.

[25] Paritosh Parmar and Brendan Tran Morris. What and how well youperformed? a multitask learning approach to action quality assessment.In Proceedings of the IEEE Conference on Computer Vision and PatternRecognition, pages 304–313, 2019.

[26] Paritosh Parmar and Brendan Tran Morris. Learning to score olympicevents. In proceedings of the IEEE Conference on Computer Vision andPattern Recognition Workshops, pages 20–28, 2017.

[27] Federico Perazzi, Jordi Pont-Tuset, Brian McWilliams, Luc Van Gool,Markus Gross, and Alexander Sorkine-Hornung. A benchmark datasetand evaluation methodology for video object segmentation. In Pro-ceedings of the IEEE Conference on Computer Vision and PatternRecognition, pages 724–732, 2016.

[28] Hamed Pirsiavash, Carl Vondrick, and Antonio Torralba. Assessing thequality of actions. In European Conference on Computer Vision, pages556–571. Springer, 2014.

[29] Mikel D Rodriguez, Javed Ahmed, and Mubarak Shah. Action macha spatio-temporal maximum average correlation height filter for actionrecognition. In CVPR, volume 1, page 6, 2008.

[30] Seyed Morteza Safdarnejad, Xiaoming Liu, Lalita Udpa, Brooks Andrus,John Wood, and Dean Craven. Sports videos in the wild (svw): Avideo dataset for sports analysis. In 2015 11th IEEE InternationalConference and Workshops on Automatic Face and Gesture Recognition(FG), volume 1, pages 1–7. IEEE, 2015.

[31] Huang-Chia Shih. A survey of content-aware video analysis for sports.IEEE Transactions on Circuits and Systems for Video Technology,28(5):1212–1231, 2017.

[32] Leonid Sigal and Michael J Black. Humaneva: Synchronized videoand motion capture dataset for evaluation of articulated human motion.Brown Univertsity TR, 120, 2006.

[33] Tomas Simon, Hanbyul Joo, Iain Matthews, and Yaser Sheikh. Handkeypoint detection in single images using multiview bootstrapping. InCVPR, 2017.

[34] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. Ucf101: Adataset of 101 human actions classes from videos in the wild. arXivpreprint arXiv:1212.0402, 2012.

[35] Sabine Susstrunk, Robert Buckley, and Steve Swen. Standard rgb colorspaces. In Color and Imaging Conference, volume 1999, pages 127–134.Society for Imaging Science and Technology, 1999.

[36] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed,Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and AndrewRabinovich. Going deeper with convolutions. In Proceedings of theIEEE conference on computer vision and pattern recognition, pages 1–9, 2015.

[37] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, andZbigniew Wojna. Rethinking the inception architecture for computervision. In Proceedings of the IEEE conference on computer vision andpattern recognition, pages 2818–2826, 2016.

[38] Mary Tuck and David Riley. The theory of reasoned action: A decisiontheory of crime. In The reasoning criminal, pages 156–169. Routledge,2017.

[39] Jacco Van Sterkenburg, Annelies Knoppers, and Sonja De Leeuw. Race,ethnicity, and content analysis of the sports media: A critical reflection.Media, Culture & Society, 32(5):819–839, 2010.

[40] Limin Wang, Yuanjun Xiong, Zhe Wang, Yu Qiao, Dahua Lin, Xiaoou

Tang, and Luc Van Gool. Temporal segment networks for actionrecognition in videos. IEEE transactions on pattern analysis andmachine intelligence, 2018.

[41] Shih-En Wei, Varun Ramakrishna, Takeo Kanade, and Yaser Sheikh.Convolutional pose machines. In CVPR, 2016.

[42] Svante Wold, Kim Esbensen, and Paul Geladi. Principal componentanalysis. Chemometrics and intelligent laboratory systems, 2(1-3):37–52, 1987.

[43] Jun Xu, Tao Mei, Ting Yao, and Yong Rui. Msr-vtt: A large videodescription dataset for bridging video and language. In Proceedings ofthe IEEE conference on computer vision and pattern recognition, pages5288–5296, 2016.

[44] Sijie Yan, Yuanjun Xiong, and Dahua Lin. Spatial temporal graphconvolutional networks for skeleton-based action recognition. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018.

[45] Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. Howtransferable are features in deep neural networks? In Advances in neuralinformation processing systems, pages 3320–3328, 2014.

[46] Joe Yue-Hei Ng, Matthew Hausknecht, Sudheendra Vijayanarasimhan,Oriol Vinyals, Rajat Monga, and George Toderici. Beyond shortsnippets: Deep networks for video classification. In Proceedings ofthe IEEE conference on computer vision and pattern recognition, pages4694–4702, 2015.