The combined Poisson INMA(q) models for time series of counts X θ

The combined Poisson INMA(q) models for time series of counts
Kaizhi Yu1, Hong Zou2
1 Statistics School, Southwestern University of Finance and Economics, Chengdu, 611130, China
2 School of Economics, Southwestern University of Finance and Economics, Chengdu, 611130, China
Corresponding author email: [email protected]
Abstract. A new stationary qth-order integer-valued moving average process with Poisson
innovation is introduced based on decision random vector. Some statistical properties of the
process are established. Estimators of the parameters of the process are obtained using the method
of moments. Some numerical results of the estimators are presented to assess the performance of
moment estimators.
Keywords. Poisson; INMA; Stationary; Ergodicity; Multinomial distribution.
1. Introduction
In natural and social sciences, time series of correlated counting are met very often. Especially
in economics and medicine many interesting variables are nonnegative count data. For example:
the number of shareholders in large Finnish and Swedish stocks, big numbers even for frequently
traded stocks, the number of arrivals per week to the emergency service of the Hospital, monthly
polio incidence counts in Germany. Most of the research on count processes assume that the count
data are independent and identically distributed. However, in practice, observations may be
autocorrelated, and this may adversely affect the performance of traditional model developed
under the assumption of independence. In recent years, count data time series models have been
devised to avoid making restrictive assmptions on the distribution of the error term. Regression
models for time series count data have been proposed (Zeger, 1988; Brannas and Johansson, 1994).
On the other hand, there have been attempts to develop suitable classes of models that resemble
the structure and properties of the usual linear ARMA models. For instances, Al-Osh and Alzaid
(1988) proposed an integer-valued moving average model (INMA) for discrete data. An integervalued GARCH model was given to study overdispersed counts (Ferland et al., 2006). Silva et
al.(2009) considered the problem of forecasting in INAR(1) model. Random coefficient INAR
models were introduced by Zheng et al. (2006,2007). The Signed thinning operator was developed
by Kachour and Truquet (2011). A new stationary first-order integer-valued autoregressive
process with geometric marginal distribution based on the generalized binomial thinning was
introduced by Ristić, Nastić and Ilić (2013). In these analysis of counts, the class of integer-valued
moving average models plays an important role.
The nonnegative integer-valued moving average process of the order q ( INMAR(q) ) was introduced
by Al-Osh and Alzaid (1988). The INMA(q) process is defined by the recursion,
X t  1 t  t 1     q t  t  q   t
(1)
1
Where
 k t  t k   i 1 vi , and the {vi } designated by counting series, is a sequence of i.i.d.
 t k
bernoulli random variables with E (vi )   , independent of
 t  k . The parameters 1 ,...,  q 
(0,1). The thinning operation ‘ t ’ indicates the corresponding thinning is associated with time.
The term
i t  t and  j t  t are independent. The choice of appropriate marginal distributions.
is still problematic for getting a particular distribution of X t . To overcome these difficulties,
Weiß (2008) introduced a combined INAR(p) models by using ‘decision’ random variables. Ristić
and Nastić (2012) considered an piecewise functions for count data model. Therefore, we adopt the
similar approach to deal with the problem in INMA(q) models. In this article, we propose an
combined INMA(q) model by allowing the parameters value are vary with ‘decision’ random
vector.
The article is organized as follows: In Sections 2, we specify the model and derive some statistical
properties. Section 3 concerns unknown parameter estimation by Yule-Walker method. In Section 4,
we conduct some Monte Carlo simulations. Finally, Section 5 concludes.
2. Definition and Basic Properties of the PCINMA(q) Process
Definition 2.1 (PCINMA(q) Model). A stochastic process { X t } is said to be the Poisson combined
INMA(q) process if it satisfies the following recursive equations:
X t  Dt ,1 (  t  t 1 )    Dt ,q (  t  t  q )   t
(2)
where { t ; t } is a sequence of independent and identically distributed Poisson random variables
with parameter
 and   (0,1) . {Dt ; t  } be an i.i.d process of ‘decision’ random vector
Dt  ( Dt ,1 ,, Dt ,q ) ~ Mult (1; 1 ,, q ) , independent of { t ; t } . Moreover, the counting
series
 t  t k , k=1, … , q are independent of {Dt } and  t at time t . To our knowledge, few
efforts has been devoted to studying combined INMA models. In this article, we aim to fill in this
gap. The Definition 2.1 shows that X t is equal to
 t  t 1   t with probability 1 , or equal to
 t  t 2   t with probability 2 , … , else equal to  t  t q   t with probability q
( 1  1  2    q 1 ) .
The moments will be useful in obtaining the appropriate estimating equations for parameter
estimation.
2
Theorem 2.1. The numerical characteristics of { X t } are as follows.
(i)
 X : E ( X t )   (  1) ,
(8)
(ii)
 X2 : Var ( X t )   (1   )  ( 2  1) 2 ,
(9)
q
 2
2 2





 j j  k ,1  k  q  1

k

j  k 1

 2
(iii)  X (k ) : cov( X t , X t  k )   k , k  q
0, k  q  1


(10)
Proof.
(i) E ( X t )  E[ Dt ,1 (  t  t 1   t )]    E[ Dt , q (  t  t  q   t )]
 E ( Dt ,1 ) E (  t  t 1   t )    E ( Dt ,q ) E (  t  t  q   t )
 1 (   )    q (   )   (  1)
(ii) Moreover,
 q

Var ( X t )  Var   Dt , j (  t  t  j   t ) 
 j 1

  Var  Dt , j (  t  t  j   t )   2 cov  Dt ,i (  t  t i   t ), Dt , j (  t  t  j   t ) 
q
j 1
i j

Note that Var Dt , j (  t  t  j   t )

in first summation on the right-hand side is
 Var ( Dt , j ) E (  t  t  j   t ) 2  E 2 ( Dt , j )Var (  t  t  j   t )
  j (1   j ) E  (  t  t  j ) 2   t2  2(  t  t  j ) t 
 j2 Var (  t  t  j )  Var ( t )  2 cov(  t  t  j ,  t ) 
  j (1   j ){[ 2 E ( t2 j )   (1   ) E ( t  j )]  E ( t2 )  2 E (  t  t  j ) E ( t )}
 j2 {[ 2Var ( t  j )   (1   ) E ( t  j )]   2 }
  j (1   j )[(  1) 2  2   (1   )  ( 2  1) 2 ]   j2 [( 2  1) 2   (1   ) ]
  j (1   j )(  1) 2  2   j [ (1   )  ( 2  1) 2 ]
3
By the same arguments as above, it follws that
cov  Dt ,i (  t  t i   t ), Dt , j (  t  t  j   t ) 
 cov  Dt ,i (  t  t i ), Dt , j (  t  t  j )   cov  Dt ,i (  t  t i ), Dt , j  t 
 cov  Dt ,i t , Dt , j (  t  t  j )   cov  Dt ,i t , Dt , j  t 
 cov( Dt ,i , Dt , j ) E (  t  t i ) E (  t  t  j )  cov( Dt ,i , Dt , j ) E (  t  t i ) E ( t )
 cov( Dt ,i , Dt , j ) E ( t ) E (  t  t  j )  cov( Dt ,i , Dt , j ) E 2 ( t )
 i j ( 2  2   2   2   2 )  i j (  1) 2  2
Therefore,
 q

Var ( X t )  Var   Dt , j (  t  t  j   t ) 
 j 1

   j (1   j )(  1)2  2   j [ (1   )  ( 2  1) 2 ]  2 i j (  1)2  2
q
j 1
i j
q


 (  1) 2  2 1    j2   [ (1   )  ( 2  1) 2 ]  2(  1) 2  2  i j
j 1
i j


  (1   )  ( 2  1) 2
(iii) For 1  k  q  1 , the autocovariance function of X t is
q
 q

cov( X t , X t  k )  cov   Dt ,i (  t  t i   t ),  Dt  k , j (  t  k  t  k  j   t  k ) 
j 1
 i 1

q


 cov  Dt ,k (  t  t  k   t ),  Dt  k , j (  t  k  t  k  j   t  k ) 
j 1


  cov  Dt , j (  t  j  t  j   t ), Dt  k , j  k (  t  k  t  k ( j  k )   t  k ) 
q
j  k 1
  cov  Dt ,k (  t  t  k ), Dt  k , j  t  k  
q
j 1
 cov  D
q
j  k 1
t, j
(  t  j  t  j ), Dt  k , j  k (  t  k  t  j ) 
q
  cov( t  t  k ,  t  k ) E ( Dt ,k ) E ( Dt  k , j )
j 1
q
  cov( t  j  t  j ,  t  k  t  j ) E ( Dt , j ) E ( Dt  k , j  k )
j  k 1
4
q
  Var ( t  k )k j 
j 1
  k   
2
2
2
q
  Var (
2
j  k 1
t j
) j j  k
q

j  k 1
j
j k
2
If k  q , by using a similar approach,we get cov( X t , X t  q )   q . For k  q  1 , all the
terms,
 t  t 1 , … ,  t  t q ,  t ,  t k  t k 1 , …,  t k  t k q and  t k involved in X t
and X t  k , are mutually independent. Therefore, the autocovariance function of X t is equal to
zero for k  q  1 .
Theorem 2.2. Let X t be the process defined by the equation in (2). Then
(i)
The { X t } is a covariance stationary process.
(ii) E ( X t )  C   , k  1, 2, 3,4 for some constant C  0 .
k
Proof. (i) The first conclusion is immediate from the definition of covariance stationary process.
(ii) For k  1, it is straightforward.
For k  2, it follows that

E ( X t2 )  max E  Dt ,1 (  t  t 1   t )  , , E  Dt ,q (  t  t  q   t ) 
2
2

Note that, for j  1, 2,..., q
E  Dt , j (  t  t  j   t )   E ( Dt , j )2 E (  t  t  j   t )2
2
 E ( Dt , j ) 2  E (  t  t  j ) 2  E ( t ) 2  2 E[(  t  t  j ) t ]
  j (1   j )[ 2 (   2 )   (1   )  (   2 )  2 2 ]
  j (1   j )(  1)[(  1)  1]
Then we have ,
1
(  1)[(  1)  1]  
4
Similarly, for k  3
E ( X t2 ) 

E ( X t3 )  max E  Dt ,1 (  t  t 1   t )  ,, E  Dt ,q (  t  t  q   t ) 
3
3

Note that, for j  1, 2,..., q
5
E  Dt , j (  t  t  j   t )   E ( Dt , j )3 E (  t  t  j   t )3
3
 E ( Dt , j )3  E (  t  t  j )3  E ( t )3  3E[(  t  t  j ) 2  t ]  E[(  t  t  j ) t2 ]
  3j [ 3  3 2 (1   )(   2 )  (  3 2 (1   )   3 ) ]
 3j   3 3j {[ 2 (   2 )   (1   ) ]}  3 3j  (   2 )
  3j {[ 2 (  1  3 )  3 (1   )  3 2  6  1]   }
Where
3
 :  3  3 2   . Let max
 max(13 ,  , q3 ) , Then
3
E ( X t3 )  max
{[ 2 (  1  3 )  3 (1   )  3 2  6  1]   }   .
After some tedious calculaions, we also can show the result holds for k  4 . We skip the details.
Next, we will present ergodic theorem for stationary process { X t } . There are a variety of ergodic
theorems, differing in their assumptions and in the modes of convergence. Here the convergence is
in mean square. The next two lemmas will be useful in proving the ergodicity of the sample mean
and smple autocovariance function of { X t } .
Lemma 2.1. If {Zt } is stationary with mean
Z and autocovariance function  Z () , then as
T   , Var ( ZT )  E ( ZT   Z ) 2  0 , if  Z (T )  0 .where ZT :
1 T
 Xt .
T t 1
Proof. See Theorem 7.1.1 in Brockwell, Peter J., and Richard A. Davis (2009) .
Lemma 2.2. Suppose {Zt } is a covariance stationary process with zero mean and covariance
function
1 T
[ E ( Z t Z t  v Z t l Z t l  v )   Z2 (v)]  0 .

T  T
l 1
 Z (v) : E ( Zt v Zt ) and. If lim
Then, for any fixed, v  0, 1, 2, …,
lim
T 
1 T
E[ Rˆ Z (v )  RZ2 (v )]2  0

l 1
T
Where ˆZ (v) is the sample covariance function ˆZ (v ) :
1 T v
 Zt v Zt .
T t 1
Proof. See Theorem 5.2 in Karlin, Samuel, and Howard E. Taylor (1975) .
Lemma 2.3. Let process Yt : X t   X is a transformation of X t , then the following results
hold.
(i)
{Yt } is a covariance stationary process with zero mean;
6
(ii)
 Y (k ) : cov(Yt , Yt k )   X (k ) , k  1, 2, 3,…;
(iii) E (Yt )  C   , k  1, 2, 3,4 for some constant C  0
k
*
*
(iv) {Yt } is ergodic in autocovariance function.
Proof. The only part of this lemma that is not obvious is part (iv). The proof of properties (iv) is
follows.
To prove that {Yt } is ergodic in autocovariance function, it suffices to show that
1 T
[ E (YtYt  vYt lYt l  v )   Y2 (v)]  0 according to lemma 2.2. Thus we will discuss two

T  T
l 1
lim
cases. For simplicity in notation we define c2 : E (Yt ) , c3 : E (Yt ) , c4 : E (Yt ) ,
2
3
4
and RY (v) : E (YY
t t v ) .
Case 1. For v  0 , E (YtYt  vYt lYt l  v )  E (Yt Yt l ) ,  Y (v)   X (v)   X
2
2
2
2
2
If 1  l  q , using the schwarz inequality we get
E (Yt 2Yt 2 l )  E (| Yt 2 |2 ) E (| Yt 2 l |2 )  c4  C *
If l  q  1 , note that Yt l and Yt are irrelevant, then
E (Yt 2Yt 2 l )  E (Yt 2 ) E (Yt 2 l )  c22   X2 (v)
Therefore,
1 T
1
[ E (YtYt  vYt lYt l  v )   Y2 (v)]  [2(c4   X2 )]  0 ,for T  

T l 1
T
Case 2. For v  1 , RY (v)  [ E (YtYt  v )]   X (v)
2
2
2
If 1  l  q  v , using schwarz inequality twice,
E (YY
E (| YtYt  v |2 ) E (| Yt lYt l  v |2 )
t t  vYt l Yt l  v ) 
1
 4 E (Yt 4 ) E (Yt 4 v ) E (Yt 4 l ) E (Yt 4 l  v )  c4
If l  q  v  1 时,note that X t X t v 与 X t l X t l  v uncorrelation,thus we have
2
E (YY
t t  vYt l Yt l  v )  E ( X t X t  v ) E (Yt  l Yt  l  v )  RY (v )
Then
1 T
1
[ E (YtYt  vYt lYt l  v )   Y2 (v)]  {(2  v)[c4  RY2 (v)]}  0 , for T   .

T l 1
T
7
This proves Lemma 2.3.
Theorem 2.3. Let { X t } be a PCINMA(q) process according to Definition 2.1. Then, the
stochastic process { X t } is ergodic in the mean and autocovariance function.
Proof. For notational simplicity, we define X T (k ) :
ˆX (k ) :
1 T k
1 T k
ˆ
X

k

Yt  k Yt ,
 t k , Y ( ) : T 
T t 1
t 1
1 T k
 ( X t k  X T )( X t  X T ) . And we assume here that the sample consists of T  k
T t 1
observations on X .
(i) Note that
 X (k )  0 , for k   . From the result of lemma 2.1, we obtain
Var ( X T )  E ( X T   X ) 2  0 .
Then X T converges in probability to
 X . Therefore, the process { X t } is ergodic in the mean.
Next, we prove the { X t } is ergodic for second moment by induction.
  X . Suppose
(ii) First we prove X T (k ) 
P
P (| X T ( k )   X |  1 )  P (| X T ( k )  X |
 P (|
1
2
1  0 is given.
)  P (| X T   X |
1
2
)


1 k
X t | 1 )  P (| X T   X | 1 )

2
2
T t 1
k


1
  P ( | X t | 1 )  P (| X T   X | 1 )
2
2
T
t 1
k
E( X t )
1
1


 0 , for T   .
P
(
|
X
|
)


t
1
T
2
t 1
t 1 2 T  1
k
Using the Markov inequality,
Scince { X t } is ergodic in the mean, thus P (| X T   X |
1
2
)  0 , for T   .
  X , for T   .
Therefore, X T (k ) 
P
P
Now we prove the second result ˆ X ( k )   X ( k ) 
0 .
Consider any
  0.
P(| ˆX (k )   X (k ) |  )  P(| ˆX (k )  ˆY (k )  ˆY (k )   X (k ) |  )


 P(| ˆX (k )  ˆY (k ) | )  P(| ˆY (k )   X (k ) | )
2
2
8
Note that


P(| ˆX (k )  ˆY (k ) | )  P(| ( X T (k )  X T )( X T   X )  ( X T2   X2 ) | )
2
2


 P (| ( X T ( k )  X T )( X T   X ) | )  P (| X T2   X2 | )
4
4
Since the sample mean X T converges in probability to
 X . According to the slutsky’s theorem,
 0 , X T   X 
0 .
we get ( X T (k )  X T )( X T   X ) 
P
Then we have P (| ˆX ( k )  ˆY (k ) |

2
2
2
P
P
) 
 0 , for T   .
From the (iv) of lemma 2.3, we obtain
P
ˆY (k )   X (k )  ˆY (k )   Y (k ) 
 0 , for T   .
 0 , for T  
And consequently, P(| ˆ X (k )   X (k ) |  ) 
P
This leads to the desired conclusion.
3. Estimation of the unknown parameters
In this section we discuss approaches to the estimation of the unknown parameters. And we assume
we have T observations, X 1 ,…, X T , from a Poisson combined INMA(q) process in which, the
order parameter, q is known. One of the main interests in the literature of INMA process is to
estimate the unknown parameters. Using the sample covariance function, we get the estimators of
unknown parameters ( 1 ,…,
q ,  ,  ) through solving the following equations.
ˆ (0)  [ (1   )  ( 2  1) 2 ]  0

q
ˆ (1)  ( 2   2 2   )  0

1
j j 1

j 2



ˆ (q  1)  ( 2   2 2  )  0
q 1
q 1

ˆ (q)   2  0
q


The idea behind these estimators is that of equating population moments to sample moments and
then solving for the parameters in terms of the sample moments. These estimators are typically
called the Yule-Walker estimators. As the ˆ (k ) consistently estimates the true autocovariance
function
 (k ) (Brockwell and Davis, 1991), the Yule-Walker estimators are consistent. Following
Brockwell and Davis (2009) and Billingsley (1961), it is easy to show that under some mild
9
moment conditions the marginal mean estimator X T is asymptotically normally distributed.
4. Monte Carlo simulation study
We provide some simulations results to show the empirical performance of these estimators.
Owing to the nonlinearity, the estimator expressions of unknown parameters are quite complicated.
The aim of simulation study is to assess the finite sample performances of the moments stimators.
Consider the following model
X t  Dt ,1 ( t  t 1 )  Dt ,2 ( t  t  2 )  Dt ,3 ( t  t 3 )   t
where { t } is a sequence of i.i.d Poisson random variables with parameter
 and   (0,1) .
The random vectors ( Dt ,1 , Dt ,2 , Dt ,3 ) is multinomial distribution with parameters (1; 1 , 2 , 3 ) ,
independent of { t } . The parameters values considered are:
(model A) (1 , 2 , 3 )  (0.1, 0.2, 0.7) ,   0.3 ,   1 ,
(model B) (1 , 2 , 3 )  (0.3,0.4,0.3) ,   0.5 ,   3 ,
(model C) (1 , 2 , 3 )  (0.6,0.2,0.2) ,   0.7 ,   10 .
The length of this discrete-valued time series T is 50, 100, 400 and 700.For each realization of
these estimators, 500 independent replicates were simulated. The numerical results of the estimators
for different true values of the parameters
(1 , 2 , 3 ) ,  , and  are presented in Table 1.
Table 1 Sample mean and mean square error (in brackets) for models A, B and C.
Sample size
Model
Parameter
1
2
A


1
B
2
50
100
400
700
0.0781
(0.0316)
0.1658
(0.0251)
0.2433
(0.0236)
1.1372
(0.1256)
0.0873
(0.0084)
0.1795
(0.0193)
0.2731
(0.0074)
1.1007
(0.0632)
0.0925
(0.0047)
0.1867
(0.0052)
0.2875
(0.0032)
1.0285
(0.0358)
0.0953
(0.0017)
0.1957
(0.0035)
0.2924
(0.0021)
1.0046
(0.0067)
0.2547
(0.0412)
0.3468
(0.0308)
0.2871
(0.0136)
0.3726
(0.0158)
0.2935
(0.0087)
0.3891
(0.0072)
0.2977
(0.0032)
0.3964
(0.0023)
10


1
2
C


All the biases of
0.4625
(0.0217)
3.5672
(0.3265)
0.4789
(0.0104)
3.1741
(0.2381)
0.4836
(0.0044)
3.0522
(0.0943)
0.4976
(0.0013)
3.0075
(0.0045)
0.5648
(0.0213)
0.1732
(0.0342)
0.6347
(0.0145)
12.3627
(1.0764)
0.5783
(0.0151)
0.1847
(0.0153)
0.6824
(0.0083)
10.6538
(0.3105)
0.5867
(0.0078)
0.1907
(0.0024)
0.6947
(0.0026)
10.1486
(0.0852)
0.5973
(0.0024)
0.1976
(0.0005)
0.6953
(0.0011)
10.0024
(0.0053)
1 , 2 and  are negative, whereas the biases of  is positive. It can be
seen that as the sample size increases, the estimates seem to converge to the true parameter values.
For example, when increasing sample size T, the Bias and MSE are both converges to zero. The
reason might be that the Yule-Walker method is based on sufficient statistics. On the other hand ,
The performance of the estimators of
 are weaker than for the ones of 1 and  .
5. Conclusion
In this article, we introduce a class of self-exciting threshold integer-valued Moving average
models driven by decision random vector. Basic probabilistic and statistical properties of this class
of models are discussed. Specifically, the method of estimation under analysis is the Yule-Walker.
Their performance is compared through a simulation study. Potential issues of future research include
to extend the results for general INARMA(p, q) models including an arbitrary distribution of
binomial thinning parameter as well as autoregressive and moving average parameters. This
remains a topic of future research.
6. Acknowledgements
We are grateful to the Professor Peter X.-K. Song from University of Michigan for his helps and
suggestions. We also thank an anonymous referee for helpful suggestions. This work is supported
by the National Natural Science Foundation of China (no. 71171166 and no. 71201126), the Science
Foundation of Ministry of Education of China (no. 12XJC910001), and the Fundamental Research
Funds for the Central Universities (no. JBK150949, JBK150501, JBK120405).
7. References
[1] Al-Osh, M.A. and A.A. Alzaid (1988). Integer-valued moving average (INMA) process.
Statist Papers. 29: 281–300.
11
[2] Billingsley, P. (1961). Statistical methods in Markov chains. Ann. Math. Statist. 32: 12–40.
[3] Brännäs, K. and P. Johansson. (1994). Time series count data regression. Comm. Statist.
Theory Methods. 23: 2907–2925.
[4] Brockwell, P.J., Davis, R.A. (1991). Time Series: Theory and Methods. Springer-Verlag,
New York. 218–219.
[5] Ferland, R., Latour, A., Oraichi, D. (2006). Integer-valued GARCH processes. J. Time Ser.
Anal. 27: 923-942.
[6] Kachour, M., Truquet, L. (2011). A p-Order signed integer-valued autoregressive (SINAR(p))
model. J. Time Ser. Anal. 32: 223–236.
[7] Karlin, Samuel, and Howard E. Taylor. (1975). A first course in stochastic processes, Academic
Press, 479–480.
[8] Ristić, Miroslav M., Aleksandar S. Nastić, and Hassan S. Bakouch. (2012). Estimation in an
integer-valued autoregressive process with negative binomial marginals (NBINAR (1)). Comm.
Statist. Theory Methods. 41: 606-618.
[9] Ristić M. M., Nastić A. S., Ilić A. V. (2013). A geometric time series model with dependent
Bernoulli counting series. J. Time Ser. Anal. 34: 466-476.
[10] Silva, N., Pereira, I., and Silva, M.E. (2009). Forecasting in INAR(1) model. REVSTAT–
Statistical Journal. 7: 119–134.
[11] Weiß, C. H. (2008). The combined INAR (p) models for time series of counts. Statist. Probab.
Lett. 78: 1817-1822.
[12] Zeger, S.L. (1988). A regression model for time series of counts. Biometrika, 75, 621–629.
[13] Zheng, H., Basawa, I.V., Datta, S. (2007). First-order random coefficient integer-valued
autoregressive processes. J. Stat. Plan. Infer. 173: 212–229.
[14] Zheng, H., Basawa, I.V., Datta, S. (2006). Inference for pth-order random coefficient integervalued autoregressive processes. J. Time Ser. Anal. 27: 411–440.
12