evaluation of a portable markerless finger position capture device: accuracy of the leap motion...

24
Title: Evaluation of a portable markerless finger position capture device: accuracy of the Leap 1 Motion controller in healthy adults 2 Authors: James Y Tung 1 , Tea Lulic 2, Dave A Gonzalez 2 , Johnathan Tran 2 , Clark R Dickerson 2 , 3 Eric A Roy 2 4 5 Affiliations: 1 Department of Mechanical and Mechatronics Engineering; 2 Department of 6 Kinesiology, University of Waterloo, Waterloo, ON, N2L 3G1 7 8 Corresponding author: James Y Tung (PhD, PEng), Department of Mechanical and Mechatronics 9 Engineering, 200 University Ave. W, Waterloo, ON, N2L 3G1 Canada, Tel. 519-888-4567 10 x33445, Email. [email protected] 11 12 Short title: Leap Motion controller finger position accuracy in healthy adults 13 14 Keywords: motor control, upper limb, kinematics, motion capture 15 16 Word count (not incl figures, tables, references, and abstract): 3,257 17 18

Upload: uwaterloo

Post on 04-Dec-2023

0 views

Category:

Documents


0 download

TRANSCRIPT

Title: Evaluation of a portable markerless finger position capture device: accuracy of the Leap 1

Motion controller in healthy adults 2

Authors: James Y Tung1, Tea Lulic2, Dave A Gonzalez2, Johnathan Tran2, Clark R Dickerson2, 3

Eric A Roy2 4

5

Affiliations: 1Department of Mechanical and Mechatronics Engineering; 2Department of 6

Kinesiology, University of Waterloo, Waterloo, ON, N2L 3G1 7

8

Corresponding author: James Y Tung (PhD, PEng), Department of Mechanical and Mechatronics 9

Engineering, 200 University Ave. W, Waterloo, ON, N2L 3G1 Canada, Tel. 519-888-4567 10

x33445, Email. [email protected] 11

12

Short title: Leap Motion controller finger position accuracy in healthy adults 13

14

Keywords: motor control, upper limb, kinematics, motion capture 15

16

Word count (not incl figures, tables, references, and abstract): 3,257 17

18

Abstract 19

Although motion analysis is frequently employed in upper limb motor assessment (e.g., 20

visually-guided reaching), they are resource-intensive and limited to laboratory settings. This study 21

evaluated the reliability and accuracy of a new markerless motion capture device, the Leap Motion 22

controller, to measure finger position. Testing conditions that influence reliability and agreement 23

between the Leap and a research-grade motion capture system were examined. Nine healthy young 24

adults pointed to 1 of 15 targets on a computer screen under 2 conditions: 1) touching the target 25

(Touch) and 2) 4 cm away from the target (No-Touch). Leap data was compared to an Optotrak 26

marker attached to the index finger. Across all trials, root mean square (RMS) error of the Leap 27

system was 17.30±9.56 mm (mean±SD), sampled at 65.47±21.53 Hz. The % viable trials and mean 28

sampling rate were significantly lower in the Touch condition (44% versus 64%, p < 0.001; 29

52.02±2.93 versus 73.98±4.48 Hz, p=0.003). While linear correlations were high (horizontal: r2 = 30

0.995, vertical r2 = 0.945), the limits of agreement were large (horizontal: -22.02 to +26.80 mm, 31

vertical: -29.41 to +30.14 mm). While not as precise as more sophisticated optical motion capture 32

systems, the Leap Motion controller is sufficiently reliable for measuring motor performance in 33

pointing tasks that do not require high positional accuracy (e.g., reaction time, Fitt’s, trails, 34

bimanual coordination). 35

36

37

1.0 Introduction 38

Assessment of upper limb impairment and function is critical for diagnosis, treatment, and 39

rehabilitation of musculoskeletal and neural disorders affecting the hand and digits. Hand and digit 40

function are commonly impaired following stroke and are among the least likely to fully recover 41

with conventional rehabilitation (Schieber et al. 2009). Assessment of upper limb function using 42

clinical instruments, such as the Nine Hole Peg Test (Oxford Grice et al. 2003) and Chedoke-43

McMaster Stroke Assessment(CMSA) scale (Gowland et al. 1993) is fundamental to developing 44

rehabilitation goals, monitoring recovery, and adapting treatment based on progress. While such 45

assessments are widely used, they may not sufficiently measure small, but clinically significant 46

changes which may better infer the state of recovery (Gladstone et al. 2002). Further, these 47

assessments require valuable time by trained therapists and are typically limited to episodic 48

intervals. As a result, there may be suboptimal feedback to clinicians and patients in the 49

rehabilitation process, leading to hindered or limited recovery. 50

In contrast, commercial motion tracking systems (e.g., Vicon, Optotrak) are capable of 51

collecting quantitative three-dimensional kinematics across a wide range of tasks (e.g., pointing, 52

reaching, grasping). These systems allow generation of in-depth kinematic measures, such as limb 53

positions and velocities accurate to sub-millimeter level (Windolf et al. 2008), to assess functional 54

capabilities. For example, kinematics of reaching and pointing tasks have been used to assess upper 55

limb impairment and recovery following stroke (Levin 1996; Coderre et al. 2010; Kitago et al. 56

2013; Kamper et al. 2002). Although optical motion analysis systems (e.g., Optitrak, Optotrak, 57

Vicon) are frequently employed in research, they are expensive (ranging from $7,500 to $415,000 58

USD in recent comparisons (Thewlis et al. 2013; Carse et al. 2013)), require training, setup time 59

for affixing markers and calibration, are sensitive to lighting conditions, need a line of sight, and 60

are typically only used in dedicated laboratory settings. These limitations present major barriers to 61

the wide clinical use of data from these motion capture systems (Rosenhahn et al. 2008) and to the 62

emerging development of at-home rehabilitation technologies (Borghese et al. 2013; Slijper et al. 63

2014; Pastor et al. 2012). 64

Advances in sensor technologies have increased accessibility and portability of upper limb 65

kinematics assessments. Inertial measurement units (IMUs), incorporating combinations of 66

accelerometer, gyroscope, and magnetometer sensors, have gained increasing use to measure 67

physical activity, gait and posture, and gross limb movements (Gebruers et al. 2010). While the 68

use of IMUs are validated for measuring the duration and intensity of upper limb movements, their 69

use in capturing the quality of upper limb movements, such as kinematics, smoothness, and 70

movement type (e.g., voluntary vs. unintentional), remains limited (Lemmens et al. 2014). 71

Furthermore, IMU techniques are susceptible to artifacts due to nearby metal objects (Brodie et al. 72

2008). While instrumented gloves (Kortier et al. 2012; Simone & Kamper 2005) for the capture of 73

hand and digit kinematics exist, their use is limited by difficulties in donning and removing the 74

gloves, especially in patients with functional impairments. A number of robotic systems, such as 75

the KinARM (Scott 1999), HapticMaster (Lemmens et al. 2014), and BONES (Milot et al. 2014) 76

systems, have also been developed to facilitate delivery of rehabilitation training. While these 77

systems are capable of high-accuracy position measurements, their utility is limited by relatively 78

high costs and long setup time. 79

Low-cost markerless optical systems have been examined as potential upper limb 80

assessment tools. The Kinect sensor (Microsoft Inc., Redmond, WA), which projects and detects 81

changes in structured infrared light patterns (i.e., dot pattern) to track the movement of body 82

segments in 3D, represents a potential alternative to measure gross body movement in a clinical 83

setting (Kurillo et al. 2013; Mobini et al. 2014). However, the spatial resolution of the Kinect limits 84

the accuracy of measuring movements of small segments (Khoshelham & Elberink 85

2012)(Khoshelham & Elberink 2012), such as fingers. To improve accuracy of estimating pointing 86

position, new techniques using proximal joint (i.e., wrist, elbow) angles to calculate pointing 87

vectors are under development (Fürntratt & Neuschmied 2014)(Fürntratt & Neuschmied 2014). 88

According to company reports, the Leap Motion controller is comprised of two 89

monochromatic stereo cameras and 3 infrared (wavelength = 850 nm) LEDs to reconstruct the 90

scene in 3D and track hand and finger position (Colgan 2014). The system’s low cost (retail 91

$80USD) and minimal set-up time, coupled with initial reports of high accuracy, indicates its 92

potential for assessment of finger movement in three-dimensional space. Independent evaluations 93

using an industrial robot with a pen-like tip reported mean squared error between the tip position 94

and position measured by the Leap system was <0.2 mm in static trials, and <2.5 mm in dynamic 95

trials (Weichert et al. 2013). In an evaluation using a plastic arm model, Guna et al. found the Leap 96

system accurate to <0.5mm in static conditions (Guna et al. 2014), further indicating initial promise 97

for clinical assessment. Furthermore, a recent comparison reported a considerably lower end-to-98

end latency of the Leap (85 ms) compared to the Kinect (170 ms) (Brown et al. 2014)(Brown et 99

al. 2014). Considering the variability in human performance compared to robotic arms is well-100

documented (e.g., van der Smagt et al. 2009), an important next step prior to clinical use is to 101

evaluate the system’s validity and reliability with human participants. 102

The purpose of this study was to evaluate the Leap system for measuring finger position 103

for clinically relevant neuromotor assessments, such as reaching and pointing, in human 104

participants. Specifically, the objectives of this study were to: 1) examine testing conditions that 105

influence the system’s reliability to measure finger position, particularly the effect of touching 106

external objects (e.g., targets on computer screen), and finger location relative to the sensor’s field 107

of view (FOV) and 2) evaluate the validity and limits of agreement between the Leap and a 108

validated research-grade optical motion capture systems. It was hypothesized that reliability of the 109

Leap system to measure finger position, indicated by % of viable trials and mean sampling rate, 110

would be compromised when touching visual stimuli presented on a computer screen compared to 111

a condition where the finger was not touching the screen. We also hypothesized that reliability will 112

be high for targets located in the reported 150 degree field of view and reduced at the margins. 113

With regards to validity and limits of agreement, it was hypothesized that correlation between the 114

Leap and Optotrak systems, measured by r2, and accuracy, measured by root mean square (RMS) 115

error, would be sufficient for reach and point tasks typically used in neuromotor assessments based 116

on previously reported upper limb assessments following stroke (e.g., Coderre et al., 2010; Kamper 117

et al., 2002; Levin, 1996). 118

119

2.0 Methods 120

2.1 Participants 121

Nine healthy right hand dominant younger adults between 18 and 30 years of age (6 122

females; 3 males) participated in this study. Participants did not report any musculoskeletal or 123

neurological disorders, and provided informed consent. This study was approved by the Office of 124

Research Ethics at the University of Waterloo. 125

2.2 Apparatus 126

A desktop computer was used to display a visual stimulus chosen randomly from 15 square 127

targets via E-prime (v 1.2, Psychology Software Tools, Inc., Sharpsberg, PA, USA). Figure 1 128

depicts the target locations and dimensions of the experimental setup. An Infrared Emitting Diode 129

(Ired) was placed on the distal phalange of the index finger and an Optotrak 3020 (Northern Digital 130

Inc., Waterloo, ON, Canada) motion capture system was used to record the location of the finger. 131

A Leap Motion controller system (version 1.2.0, Leap Motion Inc., San Francisco, CA, US), 132

connected to a laptop computer (Lenovo G550, Lenovo Inc., Beijing, China) was used to measure 133

the index finger location using a custom-built C# application to collect the Leap data stream. 134

Briefly, three infrared LEDs on the controller emit a pattern of dots which are measured using two 135

IR cameras and synthesized into 3-D positional data using a proprietary algorithm (Leap Motion 136

Inc., San Francisco, CA, US). Based on recommended placement in the user guide, the Leap sensor 137

was situated 7.5 cm in front of the display monitor and 26 cm from edge of the table (see Figure 138

1), with axes aligned with the Optotrak coordinate frame. 139

2.3 Procedure 140

From an initial starting position, participants were instructed to move their finger to 1 of 141

15 targets presented on the computer monitor (Figure 1) and maintain this position. Participants 142

were asked to perform two variations of this task: 1) touching the monitor (Touch condition) and 143

2) holding their finger 4 cm away from the screen perpendicular to the monitor at the target location 144

(No-Touch condition). In the No-Touch condition, a 4 cm diameter ball was used to control the 145

distance between the finger and monitor and was removed prior to each trial. Once the participants’ 146

fingers were in the final position, three seconds of data were collected using the Optotrak (sampling 147

rate: 120 Hz) and the Leap (sampling rate range: 50 – 120 Hz) systems. Data acquisition was 148

synchronized by simultaneous pressing of mouse buttons by the researchers. While this procedure 149

is not suitable for frame-by-frame comparisons, the reported static positions averaged over a 3 150

second trial duration is largely independent of the small variations in synchronization. 151

Participants performed 3 trials at each of 15 targets and 2 conditions for a total of 90 trials. 152

Target location was randomized for each trial. The Touch/No-Touch conditions were tested in 153

blocks, with order of presentation randomized across participants. After 5 participants were tested 154

for both Touch/No-Touch conditions, it was apparent that the Touch condition produced unreliable 155

results (56% of trials with no data, see Table 1). The remaining 4 participants were tested in the 156

No- Touch condition only. 157

2.4 Data Processing 158

All data was processed using Matlab R13 (Mathworks Inc., Natick, MA, USA). Trials in 159

which the Leap system collected 80 points or less were labeled as ‘No data’. For trials with >80 160

points, mean sampling rate was calculated as the number of points collected over the 3 s trial 161

duration. For anti-aliasing purposes, Leap and Optotrak data were digitally filtered using a dual 162

low-pass Butterworth filter (3rd order), with a cut off of 10Hz. To permit digital filtering, position 163

data collected by the Leap system was resampled at a consistent 60 Hz using a piecewise cubic 164

hermite interpolating polynomial (pchip Matlab function). To compare the Leap and Optotrak 165

systems, the center target (Target 8) was chosen as the reference origin due to the target’s high 166

reliability and consistent sampling rate (see Table 1). The reference origin was calculated as the 167

mean finger position for Target 8 trials for each participant. All position data was referenced to 168

the corresponding participants’ calculated origin. 169

2.5 Analyses 170

To evaluate the influence of the Touch condition on measurement reliability, the proportion 171

of viable trials and mean sampling rates were compared to the No-Touch condition. Reliability 172

was measured by comparing the proportion (%) of viable trials (i.e., 80 or more data points) and 173

mean sampling rate. Accuracy was measured by calculating the mean RMS error of the Leap 174

position compared to the measured Optotrak position. Fisher’s exact test was used to test the null 175

hypothesis that Touch and No-Touch conditions produced the same proportion of viable trials. 176

Similarly, task-dependent difference in mean sampling rate was compared using a paired t-test. To 177

examine the influence of field of view (FOV), the proportion of viable trials from targets at the 178

Outside edges of the grid (Targets 1, 5, 6, 10, 11, and 15) were compared to targets in the Center 179

(Targets 3, 8, and 13). Fisher’s exact test was used to test the null hypothesis that the Outside and 180

Center targets produced the same proportion of viable trials. 181

Validity and limits of agreement between the Leap and Optotrak systems were examined 182

using correlation and Bland-Altman analyses. Separate analyses were conducted for the horizontal 183

and vertical axes. In the correlation analysis, a least-squares regression line was fitted (regress 184

Matlab function). The coefficient of determination (r2) was used as a measure of linear fit. Used 185

extensively to evaluate the limits of agreement between two techniques, Bland-Altman plots were 186

constructed to examine systematic bias and estimate confidence intervals of measurement error. 187

Plots were generated using the mean of the two methods as the abscissa (x-axis) and difference 188

between Leap and Optotrack positions as the ordinate (y-axis) values. The mean errors and 95% 189

confidence intervals for each plane were reported. Considering the negative impact of the Touch 190

condition on reliability, only the No Touch data were examined in the correlation and Bland-191

Altman analyses. 192

193

3.0 Results 194

Across all conditions and targets, the mean (±SD) RMS error of the Leap system was 17.30 195

(±9.56) mm, sampled at 65.47 (±21.53) Hz, with 244/450 (54.2%) viable trials. Table 1 reports 196

descriptive statistics of means and standard deviations of RMS error, sampling rate, and % viable 197

trials by condition and target. 198

3.1 Touching vs No-Touching 199

Touching the computer screen negatively impacted the proportion of trials with viable data 200

(i.e., 80 or more data points) and mean sampling rate. As hypothesized, touching the computer 201

monitor (Touch) reduced the proportion of viable trials compared to the not touching (No Touch) 202

condition (44% versus 64%, p < 0.001). Table 2 reports contingency tables of viable trials by 203

condition. Mean (±SE) sampling rate was also lower in the Touch condition compared to the No 204

Touch condition (52.02±2.93 versus 73.98±4.48 Hz, p=0.003). 205

3.2 Outside vs Center 206

Target location affected the ability to collect finger position data. The proportion of viable 207

trials was greater in the Centre locations compared to the Outside locations (78% versus 60%, 208

p=0.009). However, in comparing the Centre and Outside targets, there was no difference in mean 209

(±SE) sampling rates (65.06±1.94 versus 68.21±1.99 Hz, p = 0.263). 210

3.3 Correlation 211

Considering the influence of the Touch condition on the reliability of the Leap system, only 212

the No Touch condition results are reported in the concurrent validity (i.e., correlation) and limits 213

of agreement (i.e., Bland-Altman) analyses. To evaluate the concurrent validity of the Leap and 214

Optotrak systems, the linear correlation between measured finger positions was examined. Overall, 215

our results show high correlation and coefficient of determination in both the horizontal (r = 0.953, 216

r2 = 0.995, Fig. 2 left) and vertical (r=0.991, r2 = 0.945, Fig. 2 right) directions. 217

3.3 Bland-Altman 218

To examine the limits of agreement between the Leap and Optotrak systems, Bland-Altman 219

plots were generated for the horizontal (Fig. 3, left) and vertical (Fig. 3, right) directions. Similar 220

to the correlation analysis, only the No Touch condition data were used due to poorer reliability in 221

the Touch condition. For the horizontal axis, the mean difference between the two systems was 222

+2.39 mm (Fig. 3, left, solid line), with lower and higher 95% limits of agreement between -22.02 223

and +26.80 mm (Fig. 3, left, dashed lines). In the vertical axis, the mean difference was +0.37 mm 224

(Fig. 3, right, solid line), with limits of agreement between -29.41 and +30.14 mm (Fig. 3, right, 225

dashed lines). 226

227

4.0 Discussion 228

To improve care and facilitate recovery of the hand and its digits, such as paresis of the 229

hand and digits following stroke, there is a need to develop new portable tools for rapid clinic- and 230

home-based assessment. Towards this goal, the purpose of this study was to evaluate the utility 231

of a new markerless kinematics capture device, the Leap Motion Controller, to measure static 232

finger position in a group of healthy younger adults. Comparison of the finger positions measured 233

by the Leap system and a research-grade optoelectronic motion capture system demonstrate that 234

the Leap system: 1) is less reliable when touching visual stimuli presented on a monitor, 2) is 235

inconsistent near the ranges of the device’s field of view (FOV), 3) provides valid measures of 236

finger position indicated by high coefficients of determination (r2), and 4) has a relatively high 237

mean RMS error and limits of agreement. While not as precise as more sophisticated systems, the 238

overall reliability and accuracy of Leap system is sufficient for pointing tasks that do not require 239

high positional accuracy measures over a limited functional space and when the finger does not 240

contact an object, such as a computer screen. However, the RMS error and limits of agreement 241

suggest that the Leap system is not precise enough to measure impaired path deviations in motor-242

impaired populations. 243

As hypothesized, the reliability of the Leap system was significantly reduced when 244

touching visual stimuli and approaching the periphery of the range of the field of view, indicated 245

by a lowered proportion of viable trials and mean sampling rate. In the current study, touching 246

targets resulted in a 20% reduction in viable trials and 21% decrease in mean sampling rate 247

compared to the No Touch condition. This may be attributable to merging of the finger and screen 248

by the Leap system, making differentiation of the two impossible. This limitation restricts the 249

ability to employ assessment tasks that require tactile interaction with buttons (e.g., reaction time), 250

screens (e.g., Trail Making Test), or objects (e.g., blocks, tools). In the evaluation of reliability, 251

outer targets demonstrated 18% less viable trials compared with targets along the center line, with 252

no difference in sampling rate. This reduction may be attributed to occlusions or merging 253

associated with the position and orientation of the hands and fingers relative to the sensor. At wider 254

angles, the pointing finger was more likely to be obstructed from view by individual hand and 255

finger orientations and anatomical dimensions. 256

The Leap system demonstrated a high degree of correlation and thus concurrent validity 257

with a validated Optotrak system (r2 = 0.995 and 0.945 in horizontal and vertical axes, 258

respectively). However, the mean RMS error (17.30 mm) in the current study is considerably 259

higher compared to previous evaluations of the Leap accuracy. Weichert et al.’s study (2013) 260

reported position error was less than 0.17 mm, and Guna et al. (2014) found a maximum standard 261

deviation of 0.49 mm across a similar space. The large differences in error are likely attributable 262

to variability in human data in the current study compared to robotic actuators (van der Smagt et 263

al. 2009). While the previous evaluations document the baseline capabilities of the technology 264

under ideal conditions, the current study better reflects conditions for clinical use. 265

The limits of agreement and mean RMS error are relatively large compared to clinically-266

relevant measures of path deviation and active range of motion in visually-guided reaching tasks. 267

In stroke studies, Levin reported a mean difference of 12.5 mm in end-point accuracy comparing 268

impaired (i.e., affected) and unimpaired arms in a rapid pointing task (Levin 1996). Schaefer et 269

al. found less than 15 mm differences in constant error between healthy controls and stroke 270

survivors (Schaefer et al. 2007; Schaefer et al. 2009). Considering the limits of agreement and 271

RMS error found in the current study are greater than these reported effect sizes, we do not 272

recommend using the Leap system to measure positional accuracy in rapid pointing tasks. 273

Despite its shortcomings in positional accuracy and precision, the Leap may be suitable for 274

a subset of motor assessment tasks, such as Trail-Making or bimanual coordination, with 275

appropriate design modifications. Adapting tasks to use non-contact targets or cues, such as 276

auditory or virtual feedback, are appropriate modifications to address contacting object limitations 277

of the Leap system. An adapted Trail-Making task (Reitan 1958), for example, could 278

simultaneously display finger position and targets on a screen to assess motor planning 279

capabilities. Auditory cues, such as a metronome, may be used to cue bimanual motor coordination 280

tasks such as circle-drawing (Bangert et al. 2010). Considering the poorer accuracy of the Leap at 281

the limits of the evaluation space, we recommend designing tasks in the center of the functional 282

volume. In particular, we recommend maintaining a height of approximately 20 cm above the 283

sensor in light of the poor reliability results at lower heights. With regards to sensitivity in the 284

depth dimension, the current study is limited to examining Leap sensitivity across varying 285

positions in the vertical and horizontal planes. Sensitivity to sensor placement in the depth 286

dimension remains to be examined. 287

Given the Leap Motion costs significantly less relative to laboratory motion tracking 288

devices, the findings in this study are promising despite the reduced accuracy. As a portable tool, 289

the sensor can be used to collect data in a patient’s home, integrated as a feedback system, and as 290

a clinical rehabilitation tool. The present findings were executed on the first commercially-291

available version of the Leap Motion sensor (version 1.2.0). Since the company is working towards 292

improving the tracking capabilities of the device, the results presented in this study may differ 293

from the findings in future studies, which use later, more refined device versions. Furthermore, 294

research is needed to examine dynamic response characteristics to evaluate the Leap’s ability to 295

measure additional kinematic variables (i.e., velocities and acceleration). 296

297

Conclusion 298

The LEAP sensor successfully measured relative 2-D coordinates of the index finger with 299

the highest correlations achieved in the left-right and up-down axes in the No Touch condition, 300

using the Optotrak system as a gold standard. Overall, the Leap system’s accuracy for finger 301

location was 17.3±9.56SD mm compared to criterion standard optical motion capture system 302

(Optotrak). However, some caution is warranted considering the relatively large limits of 303

agreement (horizontal: -22.0 to 26.8 mm, vertical: -29.4 to 30.1 mm) and limitations with 304

contacting objects. Hence, Leap Motion should provide a portable 3-D motion capture system for 305

performing basic clinical assessments in tasks such as trail-making, Fitts, or simple/choice reaction 306

time. One of the major advantages of Leap sensor is its ability to generate continuous data as the 307

hands are moved in space. Thus, it is capable of collecting a greater quantity and quality of 308

information relative to the conventional physiotherapy tools. 309

310

Acknowledgements 311

We acknowledge the assistance of Leah van Draanen with collection of data for this study. The 312

authors declare no conflicts of interest in association with this study. 313

314

315

References 316

Bangert, A.S. et al., 2010. Bimanual coordination and aging: neurobehavioral implications. 317 Neuropsychologia, 48(4), pp.1165–70. 318

Borghese, N.A. et al., 2013. Computational Intelligence and Game Design for Effective At-319 Home Stroke Rehabilitation. Games for health journal, 2(2), pp.81–88. 320

Brodie, M.A., Walmsley, A. & Page, W., 2008. Dynamic accuracy of inertial measurement units 321

during simple pendulum motion. Computer methods in biomechanics and biomedical 322 engineering, 11(3), pp.235–42. 323

Brown, M.A., Stuerzlinger, W. & Filho, E.J.M., 2014. The performance of un-instrumented in-324 air pointing. In Proceedings of the 2014 Graphics Interface Conference. Canadian 325

Information Processing Society, pp. 59–66. 326

Carse, B. et al., 2013. Affordable clinical gait analysis: an assessment of the marker tracking 327 accuracy of a new low-cost optical 3D motion analysis system. Physiotherapy, 99(4), 328 pp.347–51. 329

Coderre, A.M. et al., 2010. Assessment of upper-limb sensorimotor function of subacute stroke 330 patients using visually guided reaching. Neurorehabilitation and neural repair, 24(6), 331 pp.528–41. 332

Colgan, A., 2014. How does the Leap Motion Controller work? blog.leapmotion.com, 333 p.http://blog.leapmotion.com/hardware–to–software–ho. Available at: 334

http://blog.leapmotion.com/hardware-to-software-how-does-the-leap-motion-controller-335 work/. 336

Fürntratt, H. & Neuschmied, H., 2014. Evaluating Pointing Accuracy on Kinect V2 Sensor. In 337 International Conference on Multimedia and Human-Computer Interaction (MHCI). 338

Gebruers, N. et al., 2010. Monitoring of Physical Activity After Stroke: A Systematic Review of 339

Accelerometry-Based Measures. Archives of physical medicine and rehabilitation, 91(2), 340 pp.288–297. 341

Gladstone, D.J., Danells, C.J. & Black, S.E., 2002. The Fugl-Meyer Assessment of Motor 342 Recovery after Stroke: A Critical Review of Its Measurement Properties. 343 Neurorehabilitation and Neural Repair, 16(3), pp.232–240. 344

Gowland, C. et al., 1993. Measuring physical impairment and disability with the Chedoke-345 McMaster Stroke Assessment. Stroke, 24(1), pp.58–63. 346

Guna, J. et al., 2014. An analysis of the precision and reliability of the leap motion sensor and its 347

suitability for static and dynamic tracking. Sensors, 14(2), pp.3702–20. 348

Kamper, D.G. et al., 2002. Alterations in reaching after stroke and their relation to movement 349 direction and impairment severity. Archives of Physical Medicine and Rehabilitation, 83(5), 350 pp.702–707. 351

Khoshelham, K. & Elberink, S.O., 2012. Accuracy and resolution of Kinect depth data for indoor 352

mapping applications. Sensors, 12(2), pp.1437–54. 353

Kitago, T. et al., 2013. Improvement after constraint-induced movement therapy: recovery of 354 normal motor control or task-specific compensation? Neurorehabilitation and neural 355

repair, 27(2), pp.99–109. 356

Kortier, H.G. et al., 2012. Ambulatory assesment of hand kinematics : using an instrumented 357

glove. In XII International Symposium on 3D Analysis of Human Movement. Universita di 358

Bologna, pp. 19–29. 359

Kurillo, G. et al., 2013. Upper extremity reachable workspace evaluation with Kinect. Studies in 360 health technology and informatics, 184, pp.247–53. 361

Lemmens, R.J.M. et al., 2014. Accelerometry measuring the outcome of robot-supported upper 362 limb training in chronic stroke: a randomized controlled trial. PloS one, 9(5), p.e96414. 363

Levin, M.F., 1996. Interjoint coordination during pointing movements is disrupted in spastic 364 hemiparesis. Brain, 119(1), pp.281–293. 365

Milot, M.-H. et al., 2014. Corticospinal Excitability as a Predictor of Functional Gains at the 366 Affected Upper Limb Following Robotic Training in Chronic Stroke Survivors. 367 Neurorehabilitation and neural repair, 28(9), pp.819–827. 368

Mobini, A., Behzadipour, S. & Saadat Foumani, M., 2014. Accuracy of Kinect’s skeleton 369 tracking for upper body rehabilitation applications. Disability and rehabilitation. Assistive 370 technology, 9(4), pp.344–52. 371

Oxford Grice, K. et al., 2003. Adult norms for a commercially available Nine Hole Peg Test for 372

finger dexterity. In American journal of occupational therapy. pp. 570–3. 373

Pastor, I., Hayes, H.A. & Bamberg, S.J.M., 2012. A feasibility study of an upper limb 374 rehabilitation system using Kinect and computer games. Engineering in Medicine and 375 Biology Society (EMBC), 2012 Annual International Conference of the IEEE, 2012, 376 pp.1286–9. 377

Reitan, R.M., 1958. Validity of the Trail Making Test as an indicator of organic brain damage. 378 Perceptual and Motor Skills, 8(3), pp.271–276. 379

Rosenhahn, B., Klette, R. & Metaxas, D. eds., 2008. Human Motion, Dordrecht: Springer 380

Netherlands. 381

Schaefer, S.Y., Haaland, K.Y. & Sainburg, R.L., 2009. Dissociation of initial trajectory and final 382 position errors during visuomotor adaptation following unilateral stroke. Brain research, 383 1298, pp.78–91. 384

Schaefer, S.Y., Haaland, K.Y. & Sainburg, R.L., 2007. Ipsilesional motor deficits following 385

stroke reflect hemispheric specializations for movement control. Brain, 130(8), pp.2146–58. 386

Schieber, M.H. et al., 2009. Selective activation of human finger muscles after stroke or 387 amputation. Advances in experimental medicine and biology, 629, pp.559–75. 388

Scott, S.H., 1999. Apparatus for measuring and perturbing shoulder and elbow joint positions 389 and torques during reaching. Journal of Neuroscience Methods, 89(2), pp.119–127. 390

Simone, L.K. & Kamper, D.G., 2005. Design considerations for a wearable monitor to measure 391 finger posture. Journal of neuroengineering and rehabilitation, 2(1), p.5. 392

Slijper, A. et al., 2014. Computer game-based upper extremity training in the home environment 393 in stroke persons: a single subject design. Journal of neuroengineering and rehabilitation, 394

11(1), p.35. 395

Van der Smagt, P. et al., 2009. Robotics of human movements. Journal of physiology, Paris, 396 103(3-5), pp.119–32. 397

Thewlis, D. et al., 2013. Next-generation low-cost motion capture systems can provide 398

comparable spatial accuracy to high-end systems. Journal of applied biomechanics, 29(1), 399 pp.112–7. 400

Weichert, F. et al., 2013. Analysis of the accuracy and robustness of the leap motion controller. 401

Sensors, 13(5), pp.6380–93. 402

Windolf, M., Götzen, N. & Morlock, M., 2008. Systematic accuracy and precision analysis of 403

video motion capturing systems--exemplified on the Vicon-460 system. Journal of 404 biomechanics, 41(12), pp.2776–80. 405

406

407

Table 1: Descriptive statistics of RMS error, sampling rate, and viable trials by condition (Touch/No Touch) and Target number. Means (SD) are reported where available.

Target

1 2 3 4 5

Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch

RMS error [mm]

79.5 (45.7)

25.5 (4.1)

23.2 (9.4)

15.4 (3.1)

11.1 (2.7)

14.8 (2.8)

15.3 (4.9)

20.0 (7.2)

58.8 (41.8)

25.0 (9.7)

sampling rate [Hz]

68.1 (25.4)

76.3 (20.1)

52.0 (28.4)

74.3 (13.8)

54.8 (18.9)

76.0 (14.3)

55.1 (23.6)

81.8 (15.7)

35.8 (10.7)

78.2 (15.1)

Viable data [% trials]

4/15 (27%)

27/27 (100%)

11/15 (73%)

26/27 (96%)

11/15 (73%)

27/27 (100%)

13/15 (87%)

27/27 (100%)

8/15 (53%)

24/27 (89%)

6 7 8* 9 10

Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch

RMS error [mm] -

24.1 (5.1)

17.8 (15.1)

12.1 (5.5)

1.0 (0.3)

4.1 (6.5)

7.1 (2.7)

7.3 (2.4) -

18.8 (10.3)

sampling rate [Hz] -

64.2 (16.0)

55.7 (16.8)

71.2 (19.4)

62.3 (13.0)

63.2 (16.7)

46.0 (13.2)

64.2 (22.2) -

66.9 (20.7)

Viable data [% trials]

0/15 (0%)

25/27 (93%)

14/15 (93%)

26/27 (96%)

15/15 (100%)

24/27 (89%)

12/15 (80%)

21/27 (78%)

0/15 (0%)

21/27 (78%)

11 12 13 14 15

Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch Touch No

Touch

RMS error [mm] 249.6 - -

27.4 (11.2)

18.4 (8.5)

18.7 (9.1) - - 266.2 -

sampling rate [Hz] 41.5 73.7 -

32.4 (18.2)

48.0 (28.8)

68.2 (20.8) 61.9 30.7 17.2 -

Viable data [% trials]

1/15 (7%)

1/27 (4%)

0/15 (0%)

11/27 (41%)

8/15 (53%)

12/27 (44%)

1/15 (7%)

1/27 (4%)

1/15 (7%)

0/27 (0%)

* indicates reference target for both systems

408

409

410

411

412

413

414

415

416

417

Table 2: Contingency tables of viable trials

Touching vs. No-Touching

Viable Data No Data % Viable

Touching 99 126 44%

No-Touching 145 80 64%

Fisher’s Exact test, p < 0.001

Outside vs Center targets

Viable Data No Data % Viable

Outside 98 64 60%

Center 63 18 78%

Fisher’s Exact test, p = 0.009

418

Figure Captions 419

Figure 1.Schematic of the experimental design. 420

Figure 2.Correlation between Optotrak and Leap systems for horizontal (left panel) and vertical 421

(right panel) finger position. 422

Figure 3.Bland-Altman plots for horizontal (left panel) and vertical (right panel) for the No-423

Touching condition. 424

425

426

427

428

429

430

Figure 1 431

432

433

434 435

436

437