what is required for data assimilation that is...

6
What is Required for Data Assimilation that is Applicable to Big Data in the Solid Earth Science? ~ Importance of Simulation-/Data-driven Data Assimilation ~ Hiromichi Nagao Earthquake Research Institute The University of Tokyo 1-1-1, Yayoi, Bunkyo-ku, 113-0032 Tokyo, Japan [email protected] Abstract—Data assimilation (DA), which integrates numerical simulation models and observation data based on the Bayesian statistics, has been spreading its application field including the solid Earth science. However, the current DA is a sort of a deductive modeling method strongly depending on given simulation models, so that it never extracts, from big data, information that is beyond a priori assumptions of the simulation models. The present tutorial paper discusses the limitation of the current DA, and indicates an orientation how to implement data- driven modeling methods on DA procedure. Sparse modeling such as lasso has a potential to realize this, although specific methods are still under investigation. Keywords—data assimilation; solid Earth science; big data; data-driven modeling; sparse modeling I. INTRODUCTION Data assimilation (DA) is a fundamental computation technique to integrate numerical simulation models and observational/experimental data. DA enables us to sequentially estimate past, current and future states as well as parameters contained in a given simulation model, by embedding continuously-obtained observation data in the running simulation [1]. An essential point to be noted is that DA regards all of the model parameters and the states as stochastic variables, and estimates not a specific value but a probability distribution function (PDF) for each of the variables. This means that DA can provide a number of scenarios with probabilities that indicate how much each of the scenarios is expected to take place. A typical example of DA application can be found in the weather forecasting; the PDF related to the predicted location of a typhoon’s or hurricane’s eye at a future time is often shown by a circle, the center and the radius of which indicate the mean and the standard deviation, respectively. This expression is powerful to visually alert citizens to hazards caused by the coming typhoon. The framework of DA currently used requires an a priori simulation model that can predict possible phenomena. In other words, DA never reproduces phenomena which the given simulation model is impossible to predict. This means that the current DA is a simulation-driven “deductive modeling method”. In the case of the weather forecasting, DA can predict the track and the magnitude of an already existing typhoon through a numerical simulation, but must fully rely on observation data to know when and where a typhoon will be born since a physical simulation model cannot describe such chaotic phenomena. One of the reasons why such simulation- driven DA is enough in the weather forecasting is that most of citizens are interested in how much a generated typhoon will grow up and where it will attack rather than when and where typhoons will appear since they are always born in the Pacific Ocean far off Japan. Since it is rather rare, in most of research fields, that possible phenomena are predictable by a numerical simulation, the framework of the current DA is to be drastically improved so that DA is acceptable in different communities. In the solid Earth science, for example, the current DA hardly predicts non-linear phenomena such as earthquake occurrences and volcanic eruptions, in which citizens are usually interested. Despite this difficulty, some studies attempted to apply DA to issues in the solid Earth science such as reproduction of the large earthquakes that repeatedly occurred in history in the Nankai Trough [2], estimation of spatiotemporal distribution of afterslips caused by large earthquakes [3][4][5], clarification of acoustic wave propagation in the atmosphere excited by large earthquakes [6], and estimation of physical parameters in a volcanic eruption [7]. Recently, Earthquake Research Institute, the University of Tokyo has started a project related to DA for the solid Earth science since 2013 in order to promote the development of DA applicable to issues in this field. When we focus on geophysical observations, on the other hand, development of large-scale observation networks such as the seismometer network Hi-net and F-net, the GNSS receiver network GEONET, the earthquake and tsunami monitoring system DONET, and usage of position data of mobile phones for disaster prevention tells that “big data era” is also coming to our field. It becomes impossible for our brains to consider physical models by processing such big data distributed in ultrahigh dimensional space and time. In order to extract, from big data, unknown phenomena that are not contained in given simulation models, a data-driven DA that makes an inductive modeling from observation data possible is required as well as the current simulation-driven DA. This approach is, of course, ambitious but unavoidable to generalize DA in various fields of science.

Upload: others

Post on 06-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

What is Required for Data Assimilation that is Applicable to Big Data in the Solid Earth Science?

~ Importance of Simulation-/Data-driven Data Assimilation ~

Hiromichi Nagao Earthquake Research Institute

The University of Tokyo 1-1-1, Yayoi, Bunkyo-ku, 113-0032 Tokyo, Japan

[email protected]

Abstract—Data assimilation (DA), which integrates numerical simulation models and observation data based on the Bayesian statistics, has been spreading its application field including the solid Earth science. However, the current DA is a sort of a deductive modeling method strongly depending on given simulation models, so that it never extracts, from big data, information that is beyond a priori assumptions of the simulation models. The present tutorial paper discusses the limitation of the current DA, and indicates an orientation how to implement data-driven modeling methods on DA procedure. Sparse modeling such as lasso has a potential to realize this, although specific methods are still under investigation.

Keywords—data assimilation; solid Earth science; big data; data-driven modeling; sparse modeling

I. INTRODUCTION Data assimilation (DA) is a fundamental computation

technique to integrate numerical simulation models and observational/experimental data. DA enables us to sequentially estimate past, current and future states as well as parameters contained in a given simulation model, by embedding continuously-obtained observation data in the running simulation [1]. An essential point to be noted is that DA regards all of the model parameters and the states as stochastic variables, and estimates not a specific value but a probability distribution function (PDF) for each of the variables. This means that DA can provide a number of scenarios with probabilities that indicate how much each of the scenarios is expected to take place. A typical example of DA application can be found in the weather forecasting; the PDF related to the predicted location of a typhoon’s or hurricane’s eye at a future time is often shown by a circle, the center and the radius of which indicate the mean and the standard deviation, respectively. This expression is powerful to visually alert citizens to hazards caused by the coming typhoon.

The framework of DA currently used requires an a priori simulation model that can predict possible phenomena. In other words, DA never reproduces phenomena which the given simulation model is impossible to predict. This means that the current DA is a simulation-driven “deductive modeling method”. In the case of the weather forecasting, DA can

predict the track and the magnitude of an already existing typhoon through a numerical simulation, but must fully rely on observation data to know when and where a typhoon will be born since a physical simulation model cannot describe such chaotic phenomena. One of the reasons why such simulation-driven DA is enough in the weather forecasting is that most of citizens are interested in how much a generated typhoon will grow up and where it will attack rather than when and where typhoons will appear since they are always born in the Pacific Ocean far off Japan. Since it is rather rare, in most of research fields, that possible phenomena are predictable by a numerical simulation, the framework of the current DA is to be drastically improved so that DA is acceptable in different communities. In the solid Earth science, for example, the current DA hardly predicts non-linear phenomena such as earthquake occurrences and volcanic eruptions, in which citizens are usually interested. Despite this difficulty, some studies attempted to apply DA to issues in the solid Earth science such as reproduction of the large earthquakes that repeatedly occurred in history in the Nankai Trough [2], estimation of spatiotemporal distribution of afterslips caused by large earthquakes [3][4][5], clarification of acoustic wave propagation in the atmosphere excited by large earthquakes [6], and estimation of physical parameters in a volcanic eruption [7]. Recently, Earthquake Research Institute, the University of Tokyo has started a project related to DA for the solid Earth science since 2013 in order to promote the development of DA applicable to issues in this field.

When we focus on geophysical observations, on the other hand, development of large-scale observation networks such as the seismometer network Hi-net and F-net, the GNSS receiver network GEONET, the earthquake and tsunami monitoring system DONET, and usage of position data of mobile phones for disaster prevention tells that “big data era” is also coming to our field. It becomes impossible for our brains to consider physical models by processing such big data distributed in ultrahigh dimensional space and time. In order to extract, from big data, unknown phenomena that are not contained in given simulation models, a data-driven DA that makes an inductive modeling from observation data possible is required as well as the current simulation-driven DA. This approach is, of course, ambitious but unavoidable to generalize DA in various fields of science.

Page 2: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

The present paper introduces the foundation of DA in Section II, its application example in the solid Earth science in Section III, and suggests an orientation how to realize simulation-/data-driven DA in Section IV.

II. FOUNDATION OF DATA ASSIMILATION Since many textbooks that introduce DA have been

published, e.g., [8], we briefly summarize, in this section, the foundation of DA for readers who are unfamiliar with DA. DA starts with the following state space model (SSM):

( )1t t t tf −= +x x v (1)

( )t t t th= +y x w , (2)

where tx and ty are state and observation vectors at time t , respectively, and tv and tw are system and observation noise vectors, respectively. tf denotes a given simulation model that describes the time evolution of the state tx , and th is an observation operator to extract quantities comparable with the observation data ty by performing on components in tx . Although a more generalized SSM includes tv and tw in the arguments of tf and th , respectively, the SSM of linear version, i.e., the pair of Eqs. (1) and (2), is sufficient for most of practical cases. Eq. (1), which is called “system model”, means that the state at time t , i.e., tx , is predictable by applying the given simulation tf to the state at previous time

1t − , i.e., 1t −x . Since a perfect simulation model that describes whole phenomena never exists, the system noise tv absorbs “modeling errors” due to the imperfectness of the given simulation model. Eq. (2), which is called “observation model”, compares the observation data ty with quantities extracted from the state tx taking the observation noise tw into consideration.

Reminding that all these variables are stochastic ones and the purpose of DA is to sequentially estimate PDFs which the variables follows, we mention here the procedure of this sequential estimation in more detail. When the “filter density” at time 1t − , i.e., ( )1 1: 1|t tp − −x y , which is a PDF of 1t −x conditional on the observation data at time from 1 to 1t − , i.e.,

1: 1t −y , are given, the “predictive density” ( )1: 1|t tp −x y can be obtained as follows:

( ) ( ) ( )1: 1 1 1 1: 1 1| | |t t t t t t tp p p d∞

− − − − −−∞= x y x x x y x . (3)

Eq. (3) exactly indicates the marginal formula with respect to a conditional PDF. Note that ( )1|t tp −x x in the right-hand side can be obtained by the given simulation tf . Then the filter density at time t can be obtained using the Bayes’ theorem with an addition of observation data ty

Fig. 1. Procedure of the sequential estimation of PDFs in DA.

( ) ( ) ( )( ) ( )

1: 11:

1: 1

| ||

| |t t t t

t t

t t t t t

p pp

p p d−

−−∞

=

y x x yx y

y x x y x . (4)

Since the denominator in the right-hand side is just a normalization constant, the numerator shows the overall feature of the filter density. The likelihood ( )|t tp y x , which means the fitness between the prediction and the observation, can be computed when the PDF which the observation noise

tw follows is defined.

Fig. 1 illustrates the procedure of this alternative estimation of the predictive and filter densities. What is the most essential in the process is the selection of a sequential Bayesian filter. The Kalman filter, which estimates mean vector and covariance matrix of each PDF, is applicable if the SSM is linear, both tf and th have matrix forms, and all PDFs are normal distributions. However, an ensemble-based Bayesian filter such as the ensemble Kalman filter [8] or the particle filter [9] is needed when not all of the above conditions are satisfied. In this case, a parallel computation is required because each PDF is to be approximated by a set of realizations called “ensembles” or “particles”, which take a number of computation costs when applying to practical issues. As shown in Fig. 1, the smoothing density ( )1:|t Tp x y , where T is the final time point, can be also obtained by the smoother algorithms [9] although the problem of “degeneration”, which means that insufficient number of particles cannot approximate a PDF, easily arises especially in the case of the particle filter.

III. APPLICATION EXAMPLE OF DATA ASSIMILATION Here we introduce an application example of DA to extract

displacements of the Earth’s crust due to large earthquakes that occurred at the boundary of oceanic plates. GEONET, which is the dense array of more than 1,200 GNSS receivers established in the middle of 1990’s in Japan, has revealed the spatial and temporal distributions of activities of the Earth’s crust. Since earthquakes that occur at oceanic plate boundaries are strongly associated with several decadal changes in the crustal activities, another geophysical data available before the establishment of

Page 3: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

Fig. 2. Distribution of the tide gauge observatories in Japan.

GEONET are necessary to understand the past crustal activities. Only tide gauge data, which record the sea level at approximately 150 observatory located along the coast of Japan (Fig. 2), contain evidence of the long-term crustal displacements.

We developed a DA algorithm that extracts vertical displacements of the Earth’s crust, which were strongly associated with subduction of oceanic plates, from monthly means of the tide gauge records of several decades [10]. The SSM has a linear form

1t t tF −= +x x v (5)

t t tH= +y x w , (6)

where F and H are matrices corresponding to simulation model and observation operator, respectively. Since the complex ocean bottom topography and the convoluted coastline strongly affect the sea level changes, it is necessary to extract features from observation data via data-driven modeling rather than applying a physical-based simulation model tf , which is difficult to be provided in this case. Previous studies, e.g., [11], found that the sea level changes of several decades consist of three components; (i) linear variation due to the vertical displacement of the crust near the observatory, (ii) annual variation, and (iii) several-years variation due to oceanic activities such as the ocean general circulation in the North Pacific Ocean. The SSM defined by the pair of Eqs. (5) and (6) can extract the variations of (i)-(iii). Taking into consideration that the trend variation sometimes

Fig. 3. Results of DA applied to the tide gauge records at (a) Ayukawa in Miyagi Prefecture and (b) Hanasaki in eastern Hokkaido.

contains sudden jumps due to large earthquake at oceanic boundaries, we adopted the particle filter for sequential estimation of states at each time step. The particle filter enables us to extract such jumps by assuming a heavy-tailed non-Gaussian distribution, e.g., Cauchy distribution, for the PDF which the system noise follows.

Fig. 3 shows the result of DA application to the tide gauge data at Ayukawa in Miyagi Prefecture, which faces to the Pacific Ocean, and Hanasaki in eastern Hokkaido. The extracted trend component at Ayukawa is confirmed to be increasing for approximately forty years, which is an apparent sea level rise due to the subducting oceanic plate. This trend change actually contains a small abrupt jump due to a large earthquake of the magnitude 7.4, which occurred off-Miyagi in June, 1978. However, a multivariate analysis on Ayukawa including nearby observatories is required in order to detect such a small jump [10]. Adoption of a non-Gaussian distribution for the system noise enables us to determine the scales of two abrupt trend jumps in Hanasaki. The amplitude of the jump in 1977 probably caused by an artificial factor is approximately 50 cm, and that in 1994 indicating the crustal

(a)

(b)

Page 4: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

displacement due to the large oceanic intraplate earthquake of the magnitude 8.2, which occurred off east Hokkaido, is approximately 14 cm. Fig. 4 shows the spatial and temporal distributions of the trend variations extracted from all tide gauge observatories by our DA procedure. The features that large trend variations on the Pacific coast, which has been affected by the oceanic plate subduction, and that small variations on the coast of Sea of Japan are clearly consistent with the overall geophysical comprehension of the oceanic plate activities.

IV. TOWARD SIMULATION-/DATA-DRIVEN DATA ASSIMILATION

As mentioned in Section I, the framework of DA currently used is a deductive modeling method much depending on given simulation models. No matter how the filtering step, i.e., Eq. (4), improves model parameters and states on the basis of observation data, only future predictions which the given simulation model a priori assumes are possible. The present

Fig. 4. Distribution of the trend variations in (a) 1970, (b) 1980, (c) 1990, (d) 2000, and (e) 2003 extracted from all tide gauge observatories in Japan. The red zones indicate the areas where the crustal displacements are large.

paper focuses on the system noise tv , which absorbs modeling errors due to imperfectness of given simulation models. An implementation of an appropriate data-driven analysis method on DA could extract valuable information beyond the scope of the simulation models from the system noise (Fig. 5). One of potential candidates to realize such data-driven analysis is the “sparse modeling”, which enables us to select essential ones from redundant models and/or observation data. Here we briefly summarize one of the sparse modeling methods “lasso” (least absolute shrinkage and selection operator), which is a linear regression method based on the L1 regularization [12].

The most remarkable benefit of lasso is that it automatically selects explanatory variables in a given redundant linear regression model

01

K

k kk

y xβ β ε=

= + + , (7)

where the error ε is assumed to follow a Gaussian distribution with mean zero and variance 2σ , i.e.,

(a)

(b)

(c)

(d)

(e)

Page 5: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

Fig. 5. Comparison between (left) the simulation-driven DA currently used and (right) the simulation-/data-driven DA, which will be needed in future.

( )20,Nε σ . (8)

The coefficients ( )1, ,k k Kβ = are determined by

( ) ( )

( )

1:

2

1: 01 1

1

ˆ arg min

0

K

N Kn n

K k kn k

K

kk

y xβ

β β β

λ β λ

= =

=

= − +

+ >

, (9)

where ( ) ( )( ) ( )1: , 1, ,n nKx y n N= are observation data and λ is

a regularization parameter. The first term in the right-hand side is the loss function, which measures the fitness between the model and the observation data, and the second term is the L1 regularization term. The Lagrange undetermined multiplier leads an equivalence condition with Eq. (9)

( ) ( )

( )

1:

2

1: 01 1

1

ˆ arg min

s.t. 0

K

N Kn n

K k kn k

K

kk

y x

t t

ββ β β

β

= =

=

= − +

≤ >

, (10)

where t is a parameter one-to-one corresponding to λ . Fig. 6(a) illustrates a geometric interpretation of Eq. (10) in the case of the number of the explanatory variables is two, i.e.,

2K = . The tangent point of the elliptical contour indicating the loss function and the square indicating the regularization term gives the solution of the coefficients kβ , which satisfy Eq. (9) or (10), e.g., 2β is determined to be zero in the case of Fig. 6(a). In the case of the number of the explanatory variables is three, i.e., 3K = , two coefficients are determined to be zero if the elliptical ball is tangent to the cube at one of its vertexes, and one coefficients is zero if the tangent point is on one of the sides. In general, more coefficients are determined to be zero as the number of explanatory variables K becomes larger because the probability that the elliptic is tangent to the cube at one of its vertexes or on one of its sides increases. The most difficult problem in lasso is in the minimization of the assessment function, i.e., Eq. (9), which contains many indifferentiable points in the regularization

Fig. 6. Geometric interpretation of (a) lasso and (b) ridge regression.

term. A recently proposed algorithm termed as “glmnet” can rapidly obtain the optimum solution [13], and another study proposed a method to determine the regularization parameter λ by using an information criterion [14], although these issues are still in controversial. For comparison, Fig. 6(b) illustrates the ridge regression, in which the regularization term is replaced with the L2 norm. The ridge regression is easy to use because the assessment function has an analytical solution, but another condition or threshold is required for the selection of explanatory variables.

When the Gaussian distribution which the observation noise follows is to be estimated, it is sufficient to replace the loss function with the log-likelihood the sign of which is inversed, i.e.,

( ) ( )

2

21:

2

021 1

1

ˆ arg min log 22

1 2

K

N Kn n

k kn k

K

kk

N

y x

, σβ πσ

β βσ

λ β

= =

=

=

+ − +

+

β

. (11)

Moreover, this can be extended to the case of multidimensional observation

( ) ( )

( ) ( )

1:K

2

1:,

01 1

10

1

0

ˆ arg min log 2 log2 2

1 2

KR

TN Kn nT

k kn k

Kn nT

k kk

K

kk

lN N R

R

π

λ

= =

=

=

= +

+ − + × − +

+

y x

y x

ββ

β β

β β

β

, (12)

where l is the dimension of observation, R is the covariance matrix of the normal distribution which the observation noise follows,

2 denotes the L2 norm, and T means matrix

transposition. The form of the regularization term means

(a) (b)

Page 6: What is Required for Data Assimilation that is …confcats_isif.s3.amazonaws.com/web-files/event/...for disaster prevention tells that “big data era” is also coming to our field

“group lasso”, which can determine whether all components in kβ are zeroes (the explanatory variable kx is not needed) or

not ( kx is needed).

As mentioned in Section III, the system noise tv obtained by the current DA may include information which a given simulation model is hard to explain. In the case of the tide gauge analysis, for example, the system noise at an observatory strongly correlated with those at nearby observatories. This correlation is due to the oceanic activities, but the univariate analysis mentioned in Section III never extracts such correlation. A linear regression model that is considered to extract this effect from the obtained system noise is:

( )2, 0 ' , '

' 1 0,

Js st j j t j

jv x Nβ β ε ε σ

=

= + + , (13)

where ,st jv is the system noise at the j-th observatory

belonging to the i-th sea division, and ,1:st jx is the oceanic

activities determined by our DA procedure. The lasso is expected to select the explanatory variables related to the same sea division. The result of this analysis is expected to be almost the same with the multivariate version of the tide gauge analysis, which we carried out in the previous study [10].

V. CONCLUSIONS The present paper pointed out that the current DA, which

is a simulation-driven deductive modeling method, is insufficient to deal with big data in the solid Earth science, and suggested implementing a function of data-driven modeling method so that DA can automatically extract information, which given simulation model is hard to explain, from large amount of observation data. Sparse modeling such as lasso to extract such information from the system noise is considered to be a powerful candidate to realize this although specific methods are still under investigation.

ACKNOWLEDGMENTS The author appreciates the Geographical Survey Institute,

Japan for supplying the monthly means of tide gauge records. The map shown in this paper was generated by the Generic Mapping Tools (GMT). This work was supported by JSPS

Grant-in-Aid for Scientific Research on Innovative Areas (Grant Number 26120510).

REFERENCES [1] T. Higuchi, “Embedding reality in a numerical simulation with data

assimilation”, Proceedings of 14th International Conference on Information Fusion (FUSION), pp. 1-7, 2011.

[2] T. Hori, “Mechanisms of separation of rupture area and variation in time interval and size of great earthquakes along the Nankai Trough, southwest Japan”, J. Earth Simulator, 5, pp. 8-19, 2006.

[3] S. Miyazaki, P. Segall, J. Fukuda, and T. Kato, “Space time distribution of afterslip following the 2003 Tokachi-oki earthquake: Implications for variations in fault zone frictional properties”, Geophys. Res. Lett., 31, L06623, doi:10.102, 2004.

[4] M. Kano, S. Miyazaki, K. Ito, and K. Hirahara, “An adjoint data assimilation method for optimizing frictional parameters on the afterslip area”, Earth Planets Space, 65, No. 12, pp. 1575-1580, doi:10.5047/eps.2013.08.002, 2013.

[5] J. Fukuda, A. Kato, N. Kato, and Y. Aoki, “Are the frictional properties of creeping faults persistent? Evidence from rapid afterslip following the 2011 Tohoku-oki earthquake”, Geophys. Res. Lett., 40, pp. 3613–3617, doi:10.1002/grl.50713, 2013.

[6] H. Nagao, N. Kobayashi, S. Nakano, and T. Higuchi, “Fault parameter estimation with data assimilation on infrasound variations due to big earthquakes”, Proceedings of 14th International Conference on Information Fusion (FUSION), pp. 1-6, 2011.

[7] H. Nagao and T. Higuchi, “Data assimilation system for seismoacoustic waves”, Proceedings of 16th International Conference on Information Fusion (FUSION), pp. 1372-1377, 2013.

[8] G. Evensen, “The Ensemble Kalman Filter: theoretical formulation and practical implementation”, Ocean Dynamics, 53, pp. 343–367, doi: 10.1007/s10236-003-0036-9, 2003.

[9] G. Kitagawa, “On Monte Carlo filter and smother”, Proceedings of the Institute of Statistical Mathematics, 44, No. 1, 31-48, 1996 (in Japanese).

[10] H. Nagao, T. Higuchi, S. Miura, and D. Inazu, “Time-series modeling of tide gauge records for monitoring of the crustal activities related to oceanic trench earthquakes around Japan”, The Computer Journal, 56, No. 3, pp. 355-364, doi:10.1093/comjnl/bxs139, 2013.

[11] T. Kato and K. Tsumura, “Vertical crustal deformation in Japan deduced from tide gauge records (1951–1978)”, Bull. Earthq. Res. Inst., 54, pp. 559–628, 1979 (in Japanese).

[12] R. Tibshirani, “Regression shrinkage and selection via the lasso”, J. R. Statist. Soc., B, 58, No. 1, pp. 267-288, 1996

[13] J. Friedman, T. Hastie, and R. Tibshirani, “Regularization paths for generalized linear models via coordinate descent”, Journal of Statistical Software, 33, No. 1, 2010.

[14] K. Hirose, S. Tateishi, and S. Konishi, “Tuning parameter selection in sparse regression modeling”, Comput. Stat. Data Anal., 59, pp. 28-40, 2013.