papers i max rev

Upload: luis-carlos-gonzales-rengifo

Post on 04-Jun-2018

222 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/13/2019 Papers i Max Rev

    1/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS ARISING

    FROM GROUP REPRESENTATION THEORY AND A FAMILY OF

    QUASI-BIRTH-AND-DEATH PROCESSES

    F. ALBERTO GRUNBAUM AND MANUEL D. DE LA IGLESIA

    Abstract. We consider a family of matrix valued orthogonal polynomials obtained by I. Pacha-roni and J. A. Tirao in connection with spherical functions for the pair (SU(N+1), U(N)), see [30].

    After an appropriate conjugation, we obtain a new family of matrix valued orthogonal polynomialswhere the corresponding block Jacobi matrix is stochastic and with special probabilistic properties.This gives a highly nontrivial example of a nonhomogeneous quasi-birth-and-death process for whichwe can explicitly compute its n-step transition probability matrix and its invariant distribution.The richness of the mathematical structures involved here allows us to give these explicit results fora several parameter family of quasi-birth-and-death processes with an arbitrary (finite) number ofphases. Some of these results are plotted to show the effect that choices of the parameter values haveon the invariant distribution.

    Key words. Matrix valued orthogonal polynomials, Markov chains, block tridiagonal transitionmatrix, quasi-birth-and-death processes

    AMS subject classifications. 60J10, 42C05

    1. Purpose and contents of the paper. The aim of this paper is to tie to-

    gether two subjects that have received quite a bit of attention recently. We will notgive a detailed explanation of either one of them, since this would require too muchspace and it has been done properly in the literature already. Besides, since thesetwo subjects require rather different backgrounds, an ab-initio exposition would be aformidable task. To compensate for this we give a brief historical view of how thesetopics developed and then combine them at the appropriate point. The contents ofthis paper can be divided in three parts.

    A first part gives a brief account of the subjects that are going to play a role inthis paper. The introduction contains some historical developments tying the momentproblem with spectral theory and a quick look at birth-and-death processes, includingthe appearance of the appropriate orthogonal polynomials. Section 3 reviews verybriefly M. G. Krens theory of matrix valued orthogonal polynomials and discussesthe first example relevant to our considerations. Section 4 gives a minimal description

    of the class of Markov chains known as quasi-birth-and-death processes and talksabout the very natural connection between this and the previous section.

    The second part introduces the family of examples arising from group represen-tation theory that we are going to use in this paper. Section 5 gives a guide to theliterature on matrix valued spherical functions aimed at showing how the examplesdiscussed in Section 6 arose. Section 6 gives the bare-bones details of the extensivework carried out in [30] in the case of the complex projective space. Section 7 showshow to conjugate the weight matrix from [30] to obtain a family of matrix valuedorthogonal polynomials with extra probabilistic properties.

    The work of the first author is partially supported by NSF grant DMS-0603901, that of thesecond author is partially supported by D.G.E.S, ref. BFM2003-06335-C03-01, FQM-262 (Junta deAndaluca).

    Department of Mathematics. University of California, Berkeley. Berkeley, CA 94720 U.S.A.([email protected]).Departamento de Analisis Matematico. Universidad de Sevilla. Apdo (P. O. BOX) 1160. 41080

    Sevilla. Spain. ([email protected]).

    1

  • 8/13/2019 Papers i Max Rev

    2/20

    2 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    The third part concentrates on the probabilistic aspects of our family of exam-ples. Section 8 deals with a number of issues of probabilistic nature and displays thenetwork associated with our examples. In particular we find explicit expressions forthe invariant distribution. Section 9 gives graphical displays of some results obtainedfrom the exact formulas in the previous sections. The goal here is to show that byvarying the parameters afforded by the group theoretical situation one can obtainquite a range of different probabilistic behaviors. Finally, Section 10 gives a summaryof the results of the paper and the challenges that lie ahead.

    2. Introduction. The classical Hausdorff moment problem, that of determininga measured(x) in the interval [1, 1] from its moments

    n =

    11

    xnd(x)

    originates in very concrete problems at the end of the 19th century and was discussedby people like Chebyshev, Markov and Stieltjes. In the hands of H. Weyl and afew others, this showed the far reaching power of the modern theory of functionalanalysis in the early part of the 20th century. The main ingredient here is to connectthis problem with the spectral theory of a second order difference operator (builtfrom the momentsn) acting on functions defined on the non-negative integers. Themoments in question determine (up to scalars) a family of polynomials{Qn(x)}n0and these polynomials are the eigenfunctions of the second order difference operatoralluded to above. In the appropriate Hilbert space this operator is symmetric andthe problem of finding d(x) is the problem of finding selfadjoint extensions of thissymmetric operator. Under certain conditions there is a unique such extension andthus a unique solution to the moment problem we started from, but at any rate anyextension gives a measure d(x) that makes the polynomials orthogonal with respectto each other.

    To get closer to our subject we need a few more ingredients. One of them is givenin the rest of this section, and the other two in Sections 3 and 4.

    The presence of a second order difference operator acting on the space of functionsdefined on the non-negative integers, i.e. a semi-infinite tridiagonal matrix, makes itnatural to think of a very special kind of Markov chain on the space of non-negativeintegers. These are the so called birth-and-death processes where at each discrete unit

    of time a transition is allowed from statei to statej with probability Pij and we putPij = 0 if|i j| >1. The one-step transition probability matrix is given by

    (2.1) P =

    r0 p0q1 r1 p1

    q2 r2 p2. . .

    . . . . . .

    .

    We will assume that pj > 0, qj+1 > 0 and rj 0 for j 0. We also assumepj + rj + qj = 1 for j 1 and by putting p0+r0 1 we allow for the state j = 0 tobe an absorbing state (with probability 1 p0 r0). Some of these conditions can berelaxed.

    The problem here is to obtain an expression for the so called n-step transition

    probability matrix, giving the probability of going between any two states inn steps.By making use of the ideas mentioned above, i.e. by bringing in an appropriateHilbert space and applying then the spectral theorem, S. Karlin and J. McGregor,

  • 8/13/2019 Papers i Max Rev

    3/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 3

    [20], obtained a neat representation formula for the quantity of interest, as recalledbelow.

    If one introduces the polynomials{Qn(x)}n0 by the conditions Q1(x) = 0,Q0(x) = 1 and using the notation

    =

    Q0(x)Q1(x)

    ...

    ,

    one insists on the recursion relation

    P = x,

    it is possible to prove the existence of a unique measure d(x) supported in [1, 1]such that 1

    1

    Qi(x)Qj(x)d(x)

    11

    Qj(x)2d(x) = ij

    and one gets the Karlin-McGregor representation formula

    (2.2) Pn

    ij

    = 1

    1

    xnQi(x)Qj(x)d(x) 1

    1

    Qj(x)2d(x).

    If time is taken to be continuous, as it is done in other papers by S. Karlin andJ. McGregor, this formula and the matrix P, suffer only cosmetic changes.

    It is interesting to notice that this seminal paper of Karlin and McGregor refersboth to the standard text on the moment problem at the time, [33], as well as to thefact that W. Feller and H. P. McKean had already recognized the relevance of theHilbert space setup in the study of diffusion processes, see [7, 26]. One can mentionother papers, such as [8, 17, 19, 25] where similar ideas were at play.

    The last section of [20] deals with the case of a finite state space and the casewhen the non-negative integers are replaced by the set of all integers. Since one isusing a very powerful tool such as the spectral theorem it is clear that an adaptationof the ideas from birth-and-death processes will work here too. In the case of the

    integers, one is dealing with a state space with two singular points (one at each endof the line) and in this case H. Weyl and others had already found the correct tool:one replaces the spectral measure d(x) by a 2 2 non-negative matrix. The paperof Karlin and McGregor concludes with the explicit computation of this matrix in thecase of the doubly infinite random walk. The general formula is given in expression(12) of [20] for the case of discrete time and also in (6.8) of [18] for continuous time.

    The representation formula given above is of intrinsic interest: the computationof the left hand side of (2.2) for fixed i, j and arbitrary values ofn involves all theentries of (2.1). However, ifd(x) is known, then the right hand side of (2.2) gives away of computing this quantity using only a fixed number of entries of (2.1).

    The applicability of (2.2) depends to a large extend on our ability to obtain usefulexpressions for the orthogonal polynomials and the orthogonality measure associatedwith P. If one looks around in the literature one discovers that the number of cases

    where this is possible is rather small.We close this section by showing how one can compute in the case of a stochastic

    matrix P its invariant (stationary) distribution, i.e. the (unique up to scalars) row

  • 8/13/2019 Papers i Max Rev

    4/20

    4 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    vector

    = (0, 1, 2, . . . )

    such that

    P = .

    Recall that a matrixPwith non-negative entries is called stochastic if the sum of the

    elements in any row equals unity.We first obtain, using r0 +p0 = 1 that 1 = 0p0/q1. Then one proves by

    induction that fori 1 we have

    i = 0(p0p1 . . . pi1)/(q1q2 . . . q i).

    This has the consequence that

    i+1/i = pi/qi+1.

    Now for i 0 we have

    xQi(x) =piQi+1(x) + riQi(x) + qiQi1(x)

    with q0 = 0. Integrating this after multiplication byQi+1 or Qi1 gives 11

    xQi(x)Qi+1(x)d(x) =pi

    11

    Q2i+1(x)d(x) =qi+1

    11

    Q2i (x)d(x).

    Combining these two results we get that the ratio of the two integrals above is givenby the common value

    qi+1/pi= i/i+1.

    The moral of this is that the solution to P = can be computed (up to a commonmultiplicative scalar) either from the matrix P itself or from the knowledge of theintegrals

    11

    Q2i (x)d(x).

    In particular if we have an homogeneous birth-and-death process where pi = p andqi = qindependently of the value of i, we have that the components of are givenbyi = 0(p/q)i, i 0.

    3. Matrix valued orthogonal polynomials. We need two more charactersto be able to start our tale. The first one is the theory of matrix valued orthogonalpolynomials, whose bare bones development is given in two papers by M. G. Kren,[21, 22]. There is no written account of the motivation that led Kren to this theory,but one can easily see the connection with the spectral theory of difference operatorson the integers. This is very nicely discussed in the book by Ju. Berezanski, [2]. In

    fact the study of the classical second order difference operator on the integers is donein detail in [2] and may constitute the first example of the theory of M. G. Kren,where the polynomials and their orthogonality weight matrix W(x) are both explicitly

  • 8/13/2019 Papers i Max Rev

    5/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 5

    given. This is, of course, a special case of the matrix that appears in the last sectionof [20] for general values ofp and q(p + q= 1).

    We give now a brief account of Krens theory.Given a positive definite matrix valued measurable weight function W = W(x)

    with finite moments we can consider the skew symmetric bilinear form defined for anypair of matrix valued polynomial functions P(x) and Q(x) by the numerical matrix

    (P, Q) = (P, Q)W =

    RP(x)W(x)Q(x)dx,

    where Q(x) denotes the conjugate transpose ofQ(x). We define the matrix valuednorm ofP by

    (3.1) P2 = (P, P)W.

    One can also deal with a more general weight matrix W(x), see [4].This leads, using the Gram-Schmidt process, to the existence of a sequence of

    matrix valued orthogonal polynomials with nonsingular leading coefficients. Given anorthogonal sequence{Qn(x)}n0 one gets a three term recursion relation

    (3.2) xQn(x) = AnQn+1(x) + BnQn(x) + CnQn1(x)

    where An is non-singular. We will denote byL the corresponding Jacobi matrix,defined by the following block tridiagonal semi-infinite matrix

    L =

    B0 A0C1 B1 A1

    C2 B2 A2. . .

    . . . . . .

    .

    Using the notation

    =

    Q0(x)Q1(x)

    ...

    the relation (3.2) becomes

    (3.3) L =x.

    We will reserve the symbol P for the case whereL becomes a one-step transitionprobability matrix, thought of as a scalar matrix. The corresponding Markov chain(to appear in Section 4) will have a state space that is more complicated than the set{0, 1, 2, . . .}corresponding to a birth-and-death process featured in Section 2.

    In the scalar case, concrete examples of orthogonal polynomials, including explicitformulas for them as well as their orthogonality measure preceded the developmentof any general theory. Prominent examples are the Hermite, Laguerre and Jacobi

    polynomials. These examples arose from concrete problems in the eighteen and nine-teen centuries and played a fundamental role, in the hands of E. Schrodinger , in thedevelopment of quantum mechanics around 1925.

  • 8/13/2019 Papers i Max Rev

    6/20

    6 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    The situation in the matrix valued case is entirely different: the general theoryjust described above, came first. Until a few years ago, it may be that the onlynontrivial example was the one included in Ju. Berezanskis book, [2], alluded toabove and recalled below for the benefit of the reader.

    Consider the block tridiagonal matrix

    L=

    B0 IC1 B1 I

    C2 B2 I. . .

    . . . . . .

    with 2 2 blocks given as follows

    B0 =1

    2

    0 11 0

    , Bn = 0 ifn 1,

    Cn =1

    4I ifn 1,

    where I stands for the identity matrix. In this case one can compute explicitly thematrix valued polynomials{Qn(x)}n0 given by

    xQn(x) =Qn+1(x) +BnQn(x) + CnQn1(x), Q1(x) = 0, Q0(x) =I .

    One gets

    Qn(x) = 1

    2n

    Un(x) Un1(x)

    Un1(x) Un(x)

    ,

    whereUn(x) are the Chebyshev polynomials of the second kind.The orthogonality measure is read off from the identity

    4i

    11

    Qi(x) 11 x2

    1 xx 1

    Qj (x)dx= ijI.

    Proceeding as in [3, 10, 20] one obtains a Karlin-McGregor representation. We get,

    forn = 0, 1, 2, . . .

    Lnij =4i

    11

    xnQi(x) 11 x2

    1 xx 1

    Qj (x)dx,

    where Lnij stands for the (i, j) block of the matrix Ln. As is usual for birth-and-deathprocesses, the indices i, j run from 0 on.

    In this way, as noticed in [10], one can compute the entries of the powers Ln withL thought of as a pentadiagonal scalar matrix, namely

    L=

    0 12 1 012 0 0 1 0

    14 0 0 0 1

    . . .

    14 0 0 0

    . . .. . .

    . . . . . .

    . . .

    .

  • 8/13/2019 Papers i Max Rev

    7/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 7

    L is not a stochastic matrix since its rows do not add up to unity. Nevertheless,defining to be the 2 2 block diagonal matrix with ii = 2iIfor every block, weget from (3.3) that L1 = x and thus ifP = L1 and = , wehaveP =x. The scalar version ofPis now the stochastic matrix

    P=

    0 1212 0

    12 0 0

    12 0

    12 0 0 0

    12

    . . .

    12 0 0 0

    . . .

    . . . . . .

    . . . . . .

    .

    Observe that the norm ofQn, defined in (3.1), satisfies Qn2 = . This is nothingbut the example considered at the end of [20] in the special case ofp = q= 1/2.

    In the last few years a number of new families of matrix valued orthogonal poly-nomials have been computed explicitly along with their orthogonality measure. Typ-ically they are joint eigenfunctions of some fixed differential operator with matrixcoefficients. This search was initiated in [5], but nontrivial examples were not discov-ered until [6] and [13, 16]. The family of examples we will consider later is related to

    one of these examples and it is obtained by modifying the situation discussed in [30].

    4. Quasi-birth-and-death processes. Now we come to the last character ofour story. For our purposes, we consider a two dimensional Markov chain with discretetime. The state space consists of the pair of integers (i, j), i {0, 1, 2, . . .}, j {1, . . . , d}. The first component is usually called the level and the second one thephase. The one-step transition probability matrix, which we will denote (as before)by P has a block tridiagonal structure (see (4.2)). This indicates that in one unitof time a transition can change the phase without changing the level, or can changethe level (and possibly the phase) to either of the adjacent levels. The probability ofgoing in one step from state (i, j) to state (i, j) is given by the (j, j) element of theblockPi,i . Clearly in the case when the number of phases d is one we are back to thecase of an ordinary birth-and-death process. In general, these processes are known as

    (discrete time) quasi-birth-and-death processes.For a much more detailed presentation of this field, as well as its connections with

    queueing problems in network theory as well as the general field of communicationsystems the reader should consult [24, 27], as well as some of the references in [3].

    Once one has the notions introduced in the previous sections it is very naturalto connect them and to analyze these interesting Markov chains in terms of the cor-responding spectral properties of the resulting family of matrix valued orthogonalpolynomials. As pointed out before one can see the seed for this in the work of M.G.Kren, as well as in the original paper [20].

    We have identified two references where the corresponding Karlin-McGregor rep-resentation formula has been explicitly given, but there may be others since this issuch a natural extension of the scalar theory, see [3] and [10]. In the case of quasi-birth-and-death processes one replaces (2.2) by

    (4.1) Pnij =

    xnQi(x)W(x)Q

    j (x)dx

    Qj(x)W(x)Q

    j (x)dx

    1.

  • 8/13/2019 Papers i Max Rev

    8/20

    8 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    A different but related path to this circle of ideas in connection with networkmodels can be seen in [1].

    In [3] one finds some interesting examples where this representation formula iscomputed explicitly, including a new derivation of the result dealing with the case ofrandom walk on the integers. In [10] one finds an example taken, from [9], where theobservation is made that one has a stochastic matrix. The family of examples to beconsidered in the following sections is an extension of this example.

    Given the block transition probability matrix P, the problem of computing an

    invariant distribution row vector, i.e. a vector with non-negative entries i

    j

    = (0; 1; ) (01 , 02 , . . . , 0d; 11 , 12, . . . , 1d; )such that

    P =

    leads to a complicated system of equations.If

    (4.2) P =

    B0 A0C1 B1 A1

    C2 B2 A2. . .

    . . . . . .

    we have

    0B0+

    1C1 = 0

    and then forn 1,n1An1+

    nBn+ n+1Cn+1 =

    n.

    This gives, as is easy to check, the rather unpleasant expressions

    1 = 0(I B0)C11 ,

    2 = 0[(I

    B0)C

    11 (I

    B1)

    A0]C

    12 ,

    3 = 0[(I B0)C11 (I B1)C12 (I B2) A0C12 (I B2) (I B0)C11 A1]C13 .

    These formulas require that the matrices Cnbe invertible. Under these conditions, onecan derive nicer looking expressions for the invariant distribution (see [23]). There aremany issues here that we leave untouched in this analysis. For instance, the possibilityof choosing 0 with non-negative entries so that all the subsequent n will have thisproperty requires extra conditions.

    The general theory of quasi-birth-and-death processes is not restricted to the casewhen the matrices An, Bn and Cn are all square matrices of the same size and Anand Cn are nonsingular. It remains an interesting challenge to find a mathematicalsetup that can accommodate such a situation.

    Now that we have seen how to extend the Karlin-McGregor representation formula

    to the block tridiagonal case it is important to get examples where the polynomialsand the orthogonality matrix can be explicitly written down. This is the purpose ofthe next three sections.

  • 8/13/2019 Papers i Max Rev

    9/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 9

    5. Matrix valued spherical functions and matrix valued orthogonal

    polynomials. The theory of matrix valued spherical functions, initially discussedin [34] is one of the routes leading to explicit families of matrix valued orthogonalpolynomials and their orthogonality measure. The first family of examples appearedin [13, 16] in connection with G = SU(3). The size of the matrices here is alreadyarbitrary and the orthogonality matrix has a scalar factor of the form x(1 x).The extension to the case where this scalar factor can be taken to be x(1 x) forarbitrary, > 1 was undertaken in [9] in the 2 2 case without any reference togroup representation theory. Further examples of this kind are given in [14]. The roleof group representation theory in getting away from the 2 2 case can be seen in [15].Finally, [30] displays for the case ofG =SU(N+ 1) families of orthogonal polynomi-als depending on three parameters, , and k. In the special case ofk = +12 , onerecovers the results of [9].

    These orthogonal polynomials are given by properly packaged and conjugatedsets of matrix valued spherical functions. These spherical functions correspond toirreducible representations of U(N) and therefore are parameterized by partitions

    = (m1, m2, . . . , mN) ZN such that m1 m2 mN.

    In this paper, following [30], we use only one step representations given by a parti-tion

    = (m + , . . . , m + k

    , m , . . . , m Nk

    ), 1 k N 1.

    In terms of the parameters and , one has = m and = N 1. The remainingfree parameterwill determine the size of the corresponding matrix valued orthogonalpolynomials and it is independent ofN. In the next section it will be related to theparameterd appearing in [3].

    The examples that have been worked out so far indicate that the matrix valuedorthogonal polynomials that result from matrix valued spherical functions lead to ablock tridiagonal matrix that can be made into a stochastic one. This will be seen,for our family of examples, in Section 8.

    Although it is possible to obtain examples of stochastic matrices arising in a

    different fashion, see for instance [2], we are not aware of any other general schemethat would produce these desirable kind of matrices in a systematic fashion.

    6. A family of examples arising from the complex projective space. Inwhat follows we shall use Eij to denote the matrix with entry (i, j) equal 1 and 0elsewhere, where the indices i, j run from 0 on.

    Let d {1, 2, 3, . . .}, , >1 and 0 < k < + 1. From [30] we have that thedifferential operator

    D= x(1 x) d2

    dx2+ (C xU) d

    dx+ V,

    with matrix coefficients given by

    C=

    d1i=0

    ( + d i)Eii d2i=0

    (i+ 1)Ei+1,i, U=

    d1i=0

    ( + + d + i + 1)Eii,

  • 8/13/2019 Papers i Max Rev

    10/20

    10 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    V = d1i=0

    i( + k+ i + 1)Eii+d2i=0

    (d i 1)( k+ i + 1)Ei,i+1,

    admits as eigenfunctions (the conjugates of) a family of orthogonal polynomials withrespect to the d d weight function

    W(x) =x(1 x)Z(x), x [0, 1],where

    Z(x) =d1i,j=0

    d1r=0

    rir

    jd + k r 2

    d r 1 k+ r

    r

    (1 x)i+jxdr1Eij .Initially the value of k is an integer, but (with the appropriate notation) this

    can be taken to be any value in the range indicated above. Likewise, in the grouptheoretical setup of [30] one first takes and to be integers but then observes thatthis holds for and as above.

    In the language of [14],{ W, D}is a classical pair.In [28] one finds an explicit expression for a family of eigenfunctions ofD in terms

    of the matrix valued hypergeometric function 2F1 introduced in [35].In principle it is possible to obtain the coefficients for the three term recursion

    relation satisfied by any family of orthogonal polynomials whose existence is proved in[30]. We have not done this, but instead, since our goal is to obtain a family of matrixvalued orthogonal polynomials{Qn(x)}n0 with specific probabilistic properties, wemodify appropriately the classical pair in [30]. This is the goal of the next section.

    Remark: Notice that the notation in [13, 14, 15, 16, 30] and the one in [3] (whichwe follow here) are related by d = + 1.

    7. The new equivalent classical pair. Let us consider the following nonsin-gular upper triangular matrix:

    T =ij

    (1)i (j)i(1 d)i

    ( + k+j+ 1)i( k+ 1)i Eij ,

    where (a)nwill denote the Pochhammer symbol defined by (a)n = a(a+1) (a+n1)forn >0, (a)0 = 1. The purpose of choosing Tas above will be made clear below.

    Let us consider the new classical pair{W ,D}, whereW =TW T

    and

    D= T1DT =x(1 x) d2dx2

    + ( C xU) ddx

    + V .where

    C= d1i=0

    +d i + i(d i)( k+ i)

    + k+ 2i (i+ 1)(d i 1)( k+ i + 1)

    + k+ 2i + 2

    Eii+

    d2i=0

    1 + i +

    (i + 1)(d i 2)( k+ i + 2) +

    k+ 2i + 3

    (i+ 1)(d i 1)( k+ i + 1) +

    k+ 2i + 2

    Ei,i+1+

    d2i=0

    (i + 1)(d i 1)( k+ i + 1)

    + k+ 2i + 2 i(d i 1)( k+ i + 1)

    + k+ 2i + 1

    Ei+1,i,

  • 8/13/2019 Papers i Max Rev

    11/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 11

    U= d1i=0

    ( + + d + i + 1)Eii+i

  • 8/13/2019 Papers i Max Rev

    12/20

    12 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    forn 1, Cn is the upper bidiagonal matrix

    Cn =

    d1i=0

    n( +n + i)(k+ n + d 1)( + k+ n + d + i)(k+ n + d i 1)( + k+ n + 2i + 1)( + + 2n + d + i 1)2Eii+

    d2i=0

    n(d i 1)(k+ n + d 1)( k+ i + 1)( + + 2n + d + i)( + k+ n + 2i + 1)(k+ n + d i 2)2 Ei,i+1,

    and forn 0, Bn is the tridiagonal matrix

    Bn =

    d1i=0

    1 +

    n(k+n 1)(k+ n +d 1)(+ n + d 1)(+ + 2n + d + i 1)(k+ n + d i 2)2

    (n + 1)(k+ n)(k+ n + d)(+ n + d)

    ( + + 2n + d + i + 1)(k+ n + d i 1)2 + i(d i)(k+ d i 1)( k+ i)

    ( + k+ n + 2i)(k+ n + d i 1)2 (i+ 1)(d i 1)(k+ d i 2)( k+ i + 1)

    ( + k+ n + 2i + 2)(k+ n + d i 2)2

    Eii+

    d2i=0

    (d i 1)( k+ i + 1)( + k+ n + i + 1)( + + n + d + i)(k+ n + d i 1)( + + 2n + d + i)( + k+ n + 2i + 1)2 Ei,i+1+

    d2i=0

    (i+ 1)( + n + i + 1)(k+ d i 2)( + k+ n + d + i + 1)(k+ n + d i 2)( + + 2n + d + i + 1)( + k+ n + 2i + 2)2 Ei+1,i.

    Proof. The tools required to prove these formulas are implicitly included in [12]and [29] for the special case of= k = 1. The detailed proof of these results can beobtained following the program described above.

    The advantage of dealing with this equivalent classical pair{W ,D}, normalizedas above, shows up for instance in the fact that we have a stochastic matrix. Theentries ofAn, Bn and Cn are non-negative and by applying both sides of the identity(7.2) to the vector ed, setting x = 1 and using (7.1) we obtain that the sum of theentries in each row of our block tridiagonal matrix equals one.

    In [31] one finds a nice and different proof for the fact that our Jacobi matrix is

    stochastic in the special case of= k = 1.There are further advantages in dealing with the pair {W ,D}: an expression for

    the norms of the family{Qn(x)}n0 with respect to the matrix measureW, definedin (3.1), is given by the diagonal matrix

    Qn2W = (Qn, Qn)W =d1i=0

    (1)i (n + 1)(+ d)( + n + i + 1)(1 k d n)id 1

    i

    (d)( ++ d + i + 2n + 1)

    (7.3)

    (k+ d i 1)n( + + d + i + n)n( + k+i + n + 1)d( + k+ 2i + n + 1)(k)n( k+ 1)i(+ d)n Eii,

    where is the standard Gamma function. We point out that these matrix valuednorms are given in our case by diagonal matrices, a fact that will play an importantrole later on.

  • 8/13/2019 Papers i Max Rev

    13/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 13

    For the benefit of the reader we include here the expression for all the quantitiesabove in the case d = 1. The weightW and differential operatorD are given by

    W(x) =x(1 x) , D= x(1 x) d2dx2

    + ( + 1 + x(+ + 2))d

    dx.

    Hence, we have that the coefficients of the three term recursion relation and the normsof the orthogonal polynomials are given by

    An = (n + + 1)(n + + + 1)(2n + + + 1)(2n + + + 2)

    , n 0

    Bn =1 + n(n + )

    2n + + (n + 1)(n + + 1)

    2n + + + 2 , n 0

    Cn = n(n + )

    (2n + + )(2n + + + 1), n 1,

    Qn2W = (n + 1)(n + + 1)(+ 1)2

    (n + + 1)(n + + + 1)(2n + + + 1), n 0.

    The polynomials {Qn(x)}n0 are exactly the Jacobi polynomials in the interval [0, 1],normalized to satisfyQn(1) = 1.

    The cased = 2, for the special case ofk = +12 , can be found in [9].

    8. Probabilistic aspects of our family of examples. The correspondingJacobi matrix

    (8.1) P= L =

    B0 A0C1 B1 A1

    C2 B2 A2. . .

    . . . . . .

    made up from the coefficients introduced in Theorem 7.1 is stochastic, that isPe= e,where e denotes the semi-infinite column vector with all entries equal to 1. There-fore, it gives a block tridiagonal transition probability matrix depending on threeparameters,, and k .

    The Markov process that results from P is irreducibleandaperiodic. Indeed, onecan see that for any pair of states (i, j), (i, j), every entry in the (i, i)-block ofPn

    is positive ifn is large enough.Theorem 8.1. The Markov process that results fromP is never positive recur-

    rent. If1 < 0 then the process is null recurrent. If > 0 then the process istransient.

    Proof. One can use directly the Karlin-McGregor representation formula (4.1) toobtain Corollary 4.1 of [3]. The process turns out to be recurrent if and only if

    (8.2) eTj

    10

    W(x)1 x S

    10 dx

    ej =

    for some j {1, . . . , d}, where eT

    j = (0, . . . , 0, 1, 0, . . . , 0) denotes the j-th unit vec-tor and S0 =

    10W(x)dx =Q02W is the first moment. Otherwise, the process

    is transient. The explicit expression of the weight matrix in our case isW(x) =

  • 8/13/2019 Papers i Max Rev

    14/20

    14 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    x(1 x) Z(x), whereZ(x) is a matrix polynomial and a detailed look shows thatZ(1) =(+ 1)d1

    (d 1)!d1i,j=0

    Eij .

    Hence, condition (8.2) holds if and only if1< 0.Corollary 4.2 of [3] gives a necessary and sufficient condition for a process to be

    positive recurrent. This happens exactly when one of the entries of the measure Whasa jump at the point 1. But this is not true in our case given the form of W indicatedabove. So our process is never positive recurrent. This means that for1 < 0the process is null recurrent.

    All the considerations above result from explicit expressions forW(x) and theentries ofP.

    Now we come to a very delicate issue: the explicit computation of an invariant dis-tribution. As noticed in (7.3), our matrix valued orthogonal polynomials {Qn(x)}n0have the remarkable property that their norms are diagonal matrices. This providesus, for each n, with d scalars which could be used (inspired by the case of birth-and-death processes) as a way of getting an invariant row vector

    = (0; 1; )with the property that

    P = .

    Thus, we find the remarkable fact that the components of can be computed by therecipe

    n =ed(Qn2W)1, n 0,

    whereedis thed-dimensional row vector with all entries equal to 1. The fact that theprocess is never positive recurrent implies that there exists no invariant distributionsuch that e 1.

    The case of random walk on the integers with general values ofpand q(p+q= 1)treated in [20] gives (forp =q) an example of a transient process where the invariantdistribution is not unique. As mentioned above, it appears that even for values ofwhen our process is transient we still have a unique invariant distribution which canbe computed explicitly in terms of the norms of our orthogonal polynomials.

    To conclude this section we exhibit the network associated with our family ofexamples. The states of our network are labeled (as in any two dimensional situation)by two indices i = 0, 1, 2, . . .and j = 1, 2, . . . , d. However, to write down a one-steptransition probability matrix, one needs to agree on some linear order for the states.We use to convenient ordering

    (0, 1), (0, 2), , (0, d), (1, 1), (1, 2), , (1, d), (2, 1), (2, 2), , (2, d),

  • 8/13/2019 Papers i Max Rev

    15/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 15

    so that, for instance, the label 3 in the graph below refers to (0, 3), while the labeld + 2 refers to (1, 2), etc. This choice of lexicographic order in one of the componentsand then in the other one is an unpleasant feature that can not be avoided completely.

    The state space and the corresponding one-step transitions look as follows

    .

    .

    .

    .

    .

    .

    .

    .

    .

    .

    .

    .

    1 d + 1 2d + 1 3d + 1

    2 d + 2 2d + 2 3d + 2

    3 d + 3 2d + 3 3d + 3

    d 2d 3d 4d

    9. The shape of the invariant distribution. In this section we will study in

    more detail the behavior of the invariant distribution when the number of phases dis equal to two, a luxury we can afford since we have an analytic expression. In thiscase, the associated network takes the form

  • 8/13/2019 Papers i Max Rev

    16/20

    16 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    1 3 5 7

    2 4 6 8

    The invariant distribution such that P = is given by

    = (0; 1; )where n, n 0, is the 2-dimensional vector given by

    n =

    (n+++2)(n++2)(n++1)(n+1)(+2)2(n++k+2)

    k(2n+++2)n+k ,

    (k+1)(2n+++3)(n+++2)(n++1)(n+k+1)

    .

    From this explicit expression we can easily obtain several quantities. Data of specialinterest may be the initial value and the asymptotic behavior. The initial value isgiven by

    0 = (++3)(+1)(+2)(+k+2)

    1, (k+1)(++3)(+1)(k+1)

    .

    The asymptotic behavior follows using asymptotic formulas for the Gamma func-

    tion such as (z+)(z) z as|z| . Hence, we have

    limn

    n =

    (, ), if > 12 ,4

    (2k, 1 2k), if = 12 ,(0, 0), if 1< < 12 .

    In what follows we shall include plots of the two components n1 and n2 , as functions

    ofn, in some few representative cases. The general shape of both curves can changedepending on the values of the parameters , and k. A look at the role of theparametergives rise to four regions, namely1 <

  • 8/13/2019 Papers i Max Rev

    17/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 17

    0 1 2 3 4 5 6 7

    0

    2.5

    5

    7.5

    10

    12.5

    15

    1n

    2n

    Fig. 9.1. = 0.9, = 0.1, k= 0.8

    0 1 2 3 4 5 6 7

    0

    2.5

    5

    7.5

    10

    12.5

    15

    1n

    2n

    Fig. 9.2. = 2.5, = 0.1, k= 0.55

    0 1 2 3 4 5 6 7

    0.5

    0.6

    0.7

    0.8

    0.9

    1

    1.1

    1n

    2n

    Fig. 9.3. = 0.8, = 0.4, k= 0.3

    0 1 2 3 4 5 6 7

    0.35

    0.4

    0.45

    0.5

    0.55

    0.6

    0.65

    0.7

    1n

    2n

    Fig. 9.4. = 0.9, = 0.6, k= 0.2

    We observe that Figure 9.5, with k now approaching + 1, is similar to Figure9.4. In Figure 9.6 we observe the consequences ofk being very small.

    The last figures refer to the case = 1/2, where both components converge forlargen. In Figure 9.7 we observe that the initial value of the first component is alwayslower than the second component and that the initial value of the second componentis always greater than the first component. A small change of the value of has theeffect that the first component is always greater than the second component, as wecan see in Figure 9.8.

    Small perturbations on change the curvatures of the components in Figure 9.9with respect to Figure 9.8. In Figure 9.10 both components tend to a same valuewithout ever touching, a consequence of choosingk = +12 .

    The last two figures show how the situation can change for small perturbations ofand k. In Figure 9.11 both curves start from the same value and then they convergeto different limits, while in Figure 9.12 both components converge to the same limit.

    10. Concluding remarks. A block tridiagonal matrix L with non-negative en-tries and individual rows that add up to one gives rise to a quasi-birth-and-deathprocess. The explicit evaluation of

    Ln, for arbitrary n = 1, 2, 3, . . . can be greatly

    simplified by using ideas that go back to S. Karlin and J. McGregor and have beenexplicitly set forth in [3, 10]. The only major difficulty here is that of computing theweight matrix W(x). In this paper we start from a rich group theoretical situation

  • 8/13/2019 Papers i Max Rev

    18/20

    18 F. A. GRUNBAUM AND M. D. DE LA IGLESIA

    0 1 2 3 4 5 6 7

    0.2

    0.3

    0.4

    0.5

    0.6

    0.7

    0.8

    0.9

    1n

    2n

    Fig. 9.5. = 0.98, = 0.6, k= 0.3

    0 1 2 3 4 5 6 7

    0

    0.2

    0.4

    0.6

    1n

    2n

    Fig. 9.6. = 0.9, = 0.8, k= 0.05

    0 1 2 3 4 5 6 7

    0.3

    0.4

    0.5

    0.6

    0.7

    0.8

    0.9

    1n

    2n

    Fig. 9.7. = 0.92, = 0.5, k= 0.3

    0 1 2 3 4 5 6 7

    0.55

    0.6

    0.65

    0.7

    0.751

    n

    2n

    Fig. 9.8. = 0.6857, = 0.5, k= 0.3

    that yieldsW(x) as well as a matrix L of the type envisaged above. There are enoughfree parameters here to give instances of transient as well as recurrent Markov chains.It remains as an interesting challenge to find some real life applications to this largecollection of examples.

    REFERENCES

    [1] Basu, S. and Bose, N. K., Matrix Stieltjes series and network models, SIAM J. Matrix Anal.Applic. 14, No. 2 (1983) pp. 209222.

    [2] Berezanski, Ju. M., Expansions in Eigenfunctions of Selfadjoint Operators, Translations ofMathematical Monographs, vol. 17, American Mathematical Society, Rhode Island, 1968.

    [3] Dette, H., Reuther, B., Studden, W. and Zygmunt, M., Matrix measures and random walkswith a block tridiagonal transition matrix, SIAM J. Matrix Anal. Applic. 29, No. 1 (2006)pp. 117142.

    [4] Duran, A. J., On orthogonal polynomials with respect to positive definite matrix of measures,Can. J. Math. 47 (1995), pp. 88112.

    [5] Duran, A. J., Matrix inner product having a matrix simmetric second order differential oper-ator, Rocky Mountain J. Math. 27 (1997), pp. 585600.

    [6] Duran, A. J. and Grunbaum, F. A., Orthogonal matrix polynomials satisfying second order

    differential equations, Internat. Math. Research Notices, 2004: 10 (2004), pp. 461484.[7] Feller, W., On second order differential operators, Ann. of Math. 61, No. 1 (1955), pp. 90105.[8] Good, I. J., Random motion and analytic continued fractions, Math. Proc. Camb. Phil. Soc.

    54 (1958), pp. 4347.

  • 8/13/2019 Papers i Max Rev

    19/20

    MATRIX VALUED ORTHOGONAL POLYNOMIALS AND QBDs 19

    0 1 2 3 4 5 6 7

    0.5

    0.6

    0.7

    0.8

    0.9

    1n

    2n

    Fig. 9.9. = 0.4857, = 0.5, k= 0.3

    0 1 2 3 4 5 6 7

    0.45

    0.5

    0.55

    0.6

    0.65

    0.7

    0.75

    0.8

    1n

    2n

    Fig. 9.10. = 0.9, = 0.5, k= 0.25

    0 1 2 3 4 5 6 7

    0.4

    0.5

    0.6

    0.7

    0.8 1n

    2n

    Fig. 9.11. = 0.8, = 0.5, k= 0.3421

    0 1 2 3 4 5 6 7

    0.65

    0.675

    0.7

    0.725

    0.75

    0.775

    1n

    2

    n

    Fig. 9.12. = 0.7, = 0.5, k= 0.25

    [9] Grunbaum, F. A., Matrix valued Jacobi polynomials, Bull. Sci. Math. 127 (2003), pp. 207214.[10] Grunbaum, F. A., Random walks and orthogonal polynomials: some challenges, to appear in

    Probability, Geometry and Integrable Systems, MSRI Publication, volumen 55, 2007. Seealso arXiv math PR/0703375.

    [11] Grunbaum, F. A. y de la Iglesia, M. D., Matrix valued orthogonal polynomials related to

    SU(N + 1), their algebras of differential operators and the corresponding curves, Exp.Math. 16 , No. 2, (2007), pp. 189207.

    [12] Grunbaum, F. A., Pacharoni, I. y Tirao, J. A., A matrix valued solution to Bochners problem,J. Physics A: Math. Gen. 34 (2001), 1064710656.

    [13] Grunbaum, F. A., Pacharoni, I., and Tirao, J. A., Matrix valued spherical functions associatedto the complex projective plane, J. Functional Analysis 18 8 (2002), pp. 350441.

    [14] Grunbaum, F. A., Pacharoni, I., and Tirao, J. A., Matrix valued orthogonal polynomials of theJacobi type, Indag. Mathem. 14 3,4 (2003), pp. 353366.

    [15] Grunbaum F. A., Pacharoni I. and Tirao J. A., Matrix valued orthogonal polynomials of theJacobi type: The role of group representation theory, Ann. Inst. Fourier, Grenoble, 55 ,No. 6, (2005), pp. 20512068.

    [16] Grunbaum, F. A., Pacharoni, I. and Tirao, J. A., An invitation to matrix valued sphericalfunctions: Linearization of products in the case of the complex projective space P2(C),MSRI publication. Modern Signal Processing (D. Healy and D. Rockmore, editors) 46(2003), pp. 147160. See also arXiv math. RT/0202304.

    [17] Harris, T. E.,First passages and recurrence distributions, Trans. Amer. Math. Soc. 73 (1952),

    pp. 471486.[18] Ismail, M.E.H., Letessier, J., Masson, D. and Valent, G., Birth and death processes and orthog-

    onal polynomials, in Orthogonal Polynomials, P. Nevai (editor) Kluwer Acad. Publishers,1990, pp. 229255.

  • 8/13/2019 Papers i Max Rev

    20/20