Abstract
Motivated by the extended Poisson INAR(1), which allows innovations to be serially dependent, we develop a new family of binomial-mixed Poisson INAR(1) (BMP INAR(1)) processes by adding a mixed Poisson component to the innovations of the classical Poisson INAR(1) process. Due to the flexibility of the mixed Poisson component, the model includes a large class of INAR(1) processes with different transition probabilities. Moreover, it can capture some overdispersion features coming from the data while keeping the innovations serially dependent. We discuss its statistical properties, stationarity conditions and transition probabilities for different mixing densities (Exponential, Lindley). Then, we derive the maximum likelihood estimation method and its asymptotic properties for this model. Finally, we demonstrate our approach using a real data example of iceberg count data from a financial system.
Keywords: Count data time series, binomial-mixed Poisson INAR(1) models, mixed Poisson distribution, overdispersion, maximum likelihood estimation
1. Introduction
Modelling the integer-valued count time series has attracted a lot of attention over the last few years in a plethora of different scientific fields such as the social sciences, healthcare, insurance, economics and the financial industry. The standard ARMA model will inevitably introduce real-valued results, and so is not appropriate for modelling this type of data. As a result, many alternative classes of integer-valued time series models have been introduced and explored in the applied statistical literature. The Integer-valued autoregressive process of order one, abbreviated as INAR(1), was proposed by McKenzie [8] and Al-Osh and Alzaid [1] as a counterpart to the Gaussian AR(1) model for Poisson counts. This model was derived by manipulating the operation between coefficients and variables, as well as the innovation term, in such a way that the values are always integers. The relationship of coefficients and variables is defined as such that are i.i.d Bernoulli random variables with parameter α and ° denotes the binomial thinning operator. The binomial thinning is very easy to interpret, and binomial INAR(1) has the same autocorrelation structure as the standard AR(1) model and hence can be applied to fit the count data. For a general review, please see [11,12].
Later on, in order to accommodate different features exhibited by count data, for example, under-dispersion, overdispersion, probability of observing zero and different dependent structures, many research studies introduced alternative thinning operators or varied the distribution of for different needs. The case where are i.i.d geometric random variables is analysed by Ristić et al. [10], which is called NGINAR(1). Kirchner [7] introduced reproduction operators so that are i.i.d Poisson random variables to explore the relationship between Hawkes process and integer-valued time series. For further variation, random coefficients thinning is introduced so that are i.i.d Bernoulli with the parameter α being a random variable. This type of thinning operator was proposed by McKenzie [8,9] and Zheng et al. [14]; they applied this to a generalized INAR(1) model. In particular, to accommodate the overdispersion feature, one way is to change the thinning operators from binomial to other types as discussed above. Another way is to replace the innovation distribution with some other overdispersed distribution; for example, see [2]. A third approach would be to keep the structure of binomial INAR(1) but to allow the innovation terms to be serially dependent; see [13].
In this study, motivated by Weiß [13], we develop a new family of binomial-mixed Poisson INAR(1) (BMP INAR(1)) processes by adding a mixed Poisson component to the innovations term of the classical Poisson INAR(1) process. The proposed class of BMP INAR(1) processes is ideally suited for modelling heterogeneity in count time series data since, due to the mixed Poisson component which we introduce herein, it includes many members with different transition probabilities that can adequately capture different levels of overdispersion in the data while keeping the innovation as independent Poisson.
The rest paper is organized as follows. Section 2 defines the Binomial mixed Poisson INAR(1) model by adding a mixed Poisson component in the Poisson INAR(1) model. Statistical properties and the stationarity condition are derived in Section 3. Section 4 derives the distribution of the mixed Poisson component based on two different mixing density functions from the exponential family, namely the Exponential and Lindley distributions. In Section 5, maximum likelihood estimation is discussed as well as its asymptotic properties for the estimators. In Section 6, the model is fitted to financial data (iceberg count) and discuss numerical results. Finally, concluding remarks are provided in Section 7.
2. Construction of binomial mixed Poisson INAR(1)
In [13], the classical Poisson INAR(1) was extended by allowing the innovations ε to depend on the current state of the model such that where a and b are some positive constants. The innovation with this definition is separable in the sense that , where , with and . To introduce further heterogeneity while maintaining serially dependent innovations structure in this model, we extend this by allowing to be a mixed Poisson random variable.
Starting from a Poisson random variable U with parameter θ, we may obtain a large class of random variables by allowing θ to be another random variable which follows some classes of density function where φ can be a scalar or a vector; see Karlis [6]. The random variable U follows a Mixed Poisson distribution with g as a mixing density. The distribution function of U is defined as
| (1) |
We now construct our model.
Definition 2.1
The Binomial-Mixed Poisson integer-valued Autoregressive model (BMP INAR(1)) is defined by the following equations:
(2) where
° is a binomial thinning operator such that are i.i.d Bernoulli random variables with parameter
are i.i.d Poisson random variables with rate ;
is a reproduction operator such that are independent Mixed Poisson distributed with mixing density function ;
and ° are independent of each other so that and are independent of each other.
As we will see shortly, the stationarity condition for this model is simply where is the first moment of . When it comes to interpretation, this model can be seen as the evolution of a population where the binomial part indicates the survivors from the previous period, the mixed Poisson part is the total offspring and the innovation part indicates immigrants. Obviously, this model is a Markov Chain and its transition probability can be found easily once we know the mixing density . The probability mass function of is given by
| (3) |
where the expectation is taken over . In order to evaluate the expectation explicitly, it would be desirable that the random variables have an ‘additivity’ property such that density (or probability mass function) of the sum is either itself with different parameters or can be written in a closed form. Many members of the exponential family have this kind of property. In general, we let be of an exponential family form such that
| (4) |
Denote the density of the sum as , where are i.i.d random variables with density . The expectation above can be expressed as
| (5) |
The density is explicitly known in many cases, for example, it can be an Inverse Gaussian, Exponential, Gamma, Geometric, Bernoulli or Lindley. For the sake of parsimony, we use distributions with a single parameter. In other words, we assume that φ is scalar. Note that, if we let – a Dirac delta function concentrating at φ, the model will recover to the Extended Poisson INAR(1) in [13].
3. Statistical properties of BMP INAR(1)
3.1. Moments and correlation structure
We first need to derive the moments of .
Lemma 3.1
The first moment and second central moment of with density are given by
(6) where and .
Proof.
By the conditional expectation argument
Proposition 3.1
Assume . The stationary moments of is given by
(7)
Proof.
For the first moment, we have
Since the operators ° and are independent of each other, for the second central moment, we have
Let be the σ-algebra generated by the model up to time t, the covariance of the model is given by
Again by using conditional expectations, we have
Obviously, .
From the results above, it is clear that this model follows the same correlation structure as that of standard AR(1) model. Furthermore, unlike equal-dispersed Poisson INAR(1), BMP INAR(1) is in general an overdispersed model with Fisher index of dispersion
| (8) |
3.2. Existence of stationary solution
Proposition 3.2
Given that and the following infinite sequence:
(9) where is the probability generating function (p.g.f) of has a limit
Proof.
Define the increment of the sequence
By the definition of p.g.f, , the monotonicity of this function is shown by its first and second derivatives
By the definition of p.g.f, and . So , which implies is non-decreasing function. Then we have
Notice that . Hence we can conclude that Q is a monotonic decreasing function ranging from 0 to . In order words, for any and , the sequence is increasing with respect to i. Finally, .
Proposition 3.3
Let be the BMP INAR(1) model defined in Definition 2.1. If the condition and holds, then the process has a proper stationary distribution and is an ergodic Markov Chain. The stationary distribution is .
Proof.
Denote the p.g.f of and the innovation as and respectively, then can be expressed as following product:
To show the existence of the limiting distribution is equivalent to show the limit of the product as n goes to infinity is something other than 0, which means that we have to show that the series
is convergent as . The convergence of the infinite series can be shown by the ratio test
(10) Hence , from which we can infer that exists and the limiting distribution of exists. Furthermore, by the construction of , the chain is defined on a countable state space . The positivity of transition probability implies that is irreducible and aperiodic. Hence the limiting distribution is the unique stationary distribution for .
In general, as long as , so we just need to ensure the existence of the first moment to achieve the stationarity of . The infinite product is the p.g.f of the stationary distribution, which also satisfies
| (11) |
4. Distribution function of the mixed Poisson component
In order to apply maximum likelihood estimation for the statistical inference of this model, we need to derive the distribution of according to different density functions g. As mentioned before, we focus on the density g coming from the exponential family. For expository purposes, we will derive the distribution of based on exponential and Lindley densities.
4.1. Mixed by exponential density
If , then the distribution of is given by
| (12) |
which is a geometric distribution with parameter . Then, the distribution function of as well as its first and second derivatives are given by
| (13) |
Note that will recover to the NGINAR(1) in [10] if we further let . In general, the stationarity condition becomes and the probability generating function of satisfies the equation
| (14) |
We will now relax the assumption of the innovation term being Poisson and let the marginal distribution of X be a geometric random variable with parameter . Using the relationship of the p.g.f, we can infer the required distribution of Z.
Proposition 4.1
If or and the distribution of follows a mixed geometric distribution such that
(15) then the marginal distribution of X follows a distribution.
Proof.
By utilizing equation (14), we assume the X has a geometric distribution such that . Then, the probability generating function of Z has the following form:
(16)
4.2. Mixed by Lindley density
Suppose now the density is a Lindley density function. The distribution of is the so-called Poisson–Lindley distribution, see [6], which has the following probability mass function
| (17) |
Under this parameter setting, which makes the parameter φ less interpretable. So we adopt the following parameter setting for the mixing density
| (18) |
Then, . On the other hand, the additivity of is not that clear. In order to evaluate the expectation (3), we need to find out the distribution of .
Proposition 4.2
Suppose are i.i.d Lindley distributed. The density of the sum is given by
(19)
Proof.
We can prove this by inverting the Laplace transform. The Laplace transform of is
Then the Laplace transform of is simply the product of , which is
Using a binomial expansion, we have
Obviously, the density function of is the integrand except .
Then, the distribution of is given by the following proposition.
Proposition 4.3
The probability mass function of as well as its derivatives are given by
(20) where
Proof.
5. Maximum likelihood estimation and its asymptotic property
In general, the transition probability can be written down explicitly as
| (21) |
The log likelihood function is simply .
Proposition 5.1
Suppose we have a random sample . Let denote the parameters vector for the stationary BMP INAR(1) model. The maximum likelihood estimator has the following asymptotic distribution:
(22) where
(23)
(24) where . The first and second derivatives of each distribution function is given by
Proof.
From Proposition 3.3, we know that the is stationary and ergodic and its stationary distribution is characterized by the p.g.f . Then score functions and information matrix are also stationary and ergodic. Then the proof for asymptotic normality is similar to the proof of Theorem 4 in Appendix A of [3].
The expectation of information matrix can be calculated numerically by finding out unconditional distribution and joint distribution . However, this would be computational intensive when the sample size n is large. In practice, since the process is stationary and ergodic, when n is large.
To verify the asymptotic normality of the maximum likelihood estimators, we conduct a Monte Carlo experiment. This experiment is based on 2000 replications. For each replication, a time series of BMP-INAR(1) with chosen mixing density, either Exponential or Lindley, of size is generated. The parameters are set as for both mixing densities and they are estimated via the maximum likelihood method. The biases and standard errors of the estimated parameters are shown in Tables 1 and 2. We observe that the biases of the estimators are either reasonably small or decreasing with respect to the sample size n. And it is clear that the standard error is also decreasing with respect to n. Finally, in order to graphically inspect the distribution of estimators, normal quantile-quantile plots are provided in Figure 1.
Table 1.
The bias of Maximum likelihood estimators of BMP-INAR(1) model with respect to different sample size n.
| ) | n = 100 | n = 200 | n = 300 | n = 400 | n = 500 | |
|---|---|---|---|---|---|---|
| Exponential | 0.0022 | −0.0021 | 0.0019 | −0.0003 | −0.0003 | |
| ϕ | −0.0284 | −0.0104 | −0.0110 | −0.0072 | −0.0059 | |
| 0.1089 | 0.0526 | 0.0384 | 0.0366 | 0.0279 | ||
| Lindley | −0.0008 | 0.0004 | −0.0015 | −0.0020 | −0.0011 | |
| ϕ | −0.0209 | −0.0143 | −0.0085 | −0.0050 | −0.0039 | |
| 0.0387 | 0.0227 | 0.0141 | 0.0144 | 0.0101 |
Table 2.
The standard error of Maximum likelihood estimators of BMP-INAR(1) model with respect to different sample size n.
| S.E.( ) | n = 100 | n = 200 | n = 300 | n = 400 | n = 500 | |
|---|---|---|---|---|---|---|
| Exponential | 0.1303 | 0.0965 | 0.0752 | 0.0663 | 0.0576 | |
| ϕ | 0.1384 | 0.0970 | 0.0783 | 0.0670 | 0.0581 | |
| 0.3982 | 0.2858 | 0.2276 | 0.2012 | 0.1764 | ||
| Lindley | 0.1319 | 0.0991 | 0.0854 | 0.0711 | 0.0630 | |
| ϕ | 0.1432 | 0.1054 | 0.0880 | 0.0729 | 0.0661 | |
| 0.2050 | 0.1515 | 0.1166 | 0.0999 | 0.0911 |
Figure 1.
Quantile–Quantile plots for maximum likelihood estimators of BMP-INAR(1) model. The left panel shows plots for the Exponential mixing density, while the right panel depicts the plots for the Lindley mixing density.
6. Real data example: iceberg order data
The iceberg order counts concern the Deutsche Telekom shares traded in the XETRA system of Deutsche Börse, and the concrete time series gives the number of iceberg orders (for the ask side) per 20 min for 32 consecutive trading days in the first quarter of 2004. The special feature of iceberg orders is that only a small part of the order (tip of the iceberg) is visible in the order book and the main part of the order is hidden. For detail description, please see the [4,5]. This dataset is also analysed in [13], where the Extended Poisson INAR(1) is applied to fit the data.
A table of descriptive statistics, a time series, as well as the ACF and PACF plots are shown in Table 3 and Figure 2. The variance of the iceberg count is higher than its mean, which indicates the data is overdispersed. The level of dispersion is described by the Fisher index of dispersion . Evidence of the applicability of a first-order autoregressive model is indicated by the empirical ACF and PACF graphs. They illustrate a clear decay for ACF and cut-off at lag = 1 for PACF.
Table 3.
Descriptive statistics of iceberg count.
| Minimum | Maximum | Median | Mean | Variance | FI |
|---|---|---|---|---|---|
| 0 | 9 | 1 | 1.407 | 2.184 | 1.552 |
Figure 2.
Time series plot of iceberg data and its empirical ACF PACF plots. The dash line lines are the 95% confident bands by assuming the series to be a white noise process.
The likelihood function is constructed as in (21) with different (mixed by Exponential or Lindley). It is then maximized through ‘optim’ in R with ‘method = BFGS’ (quasi-Newton method) while the standard deviations of MLEs are calculated through inverting the negative observed information matrix in Proposition 5.1 based on MLEs. To access the goodness of fit, we adopt the information criteria AIC and BIC as well as the (standardized) Pearson residuals. If the model is correctly specified, Pearson residuals for BMP-INAR(1) are expected to have a mean and variance close to 0 and 1, respectively, with no significant autocorrelation. The Pearson residuals are calculated by the following formula:
| (25) |
where denotes the observed value.
The ACF plots of the Pearson residuals in Figure 3 indicate that the BMP-INAR(1) models are appropriate for fitting the iceberg data. The estimated parameters shown in Table 4 are significantly different from 0, which is indicated by their estimated standard deviation. Compared to the Dirac delta case, which is actually the Extended Poisson INAR(1) of [13], the other two cases do show some improvement with smaller AIC, BIC values and larger fitted Fisher index of dispersion which, however, is slightly smaller than the empirical . On the other hand, it seems that there is little difference between the other two cases as they have very similar AIC and BIC values. This is due to the fact that the value of is identical for both densities. Finally, it should be noted that the variance of the Pearson residuals is visibly larger than 1. As it was previously mentioned, the exponential and Lindley mixing densities were considered for expository purposes. Therefore, since the proposed family of BMP INAR(1) models is quite general, another mixing distribution could potentially more efficiently capture the observed dispersion structure for this data.
Figure 3.
Autocorrelation of Standardized Pearson residuals for three different mixing densities.
Table 4.
The results for the BMP INAR(1) model mixed by different density functions.
| Pearson residuals | ||||||||
|---|---|---|---|---|---|---|---|---|
| Mixing density | AIC | BIC | Mean | Variance | ||||
| Dirac delta | 0.410 | 0.188 | 0.567 | 2212 | 2226 | −0.001 | 1.159 | 1.295 |
| (0.058) | (0.059) | (0.040) | ||||||
| Exponential | 0.434 | 0.167 | 0.563 | 2208 | 2222 | −0.002 | 1.154 | 1.315 |
| (0.044) | (0.044) | (0.040) | ||||||
| Lindley | 0.434 | 0.167 | 0.563 | 2208 | 2222 | −0.002 | 1.154 | 1.314 |
| (0.043) | (0.043) | (0.040) | ||||||
Note: The results of Dirac delta case are from Table 2 of [13]. The estimated standard deviations for all models are in brackets.
Overall, the mixed Poison component in the BMP INAR(1) model efficiently captures the overdispersion in this type of financial data.
7. Concluding remarks
The BMP INAR(1) is an extension of the classical Poisson INAR(1) model obtained by adding an additional mixed Poisson component and hence it can capture the level of overdispersion coming from the data. The exponential family is a desired choice for the mixing density due to its ‘additivity’ property. The choice of the mixing density can control the dispersion level to some extent, although the BMP INAR(1) is always overdispersed in general. Furthermore, due to its simplicity, is actually a Markov chain and the maximum likelihood estimation method can be applied easily. The real data analysis shows that BMP INAR(1) can be a potential choice for modelling financial count data that exhibit standard AR(1) structure and overdispersion.
Acknowledgments
The authors would like to thank the anonymous referee for their very helpful comments and suggestions which have significantly improved this article and would like to thank Prof Christian Wei for kindly sharing the financial count data.
Disclosure statement
No potential conflict of interest was reported by the author(s).
References
- 1.Al-Osh M. and Alzaid A.A., First-order integer-valued autoregressive (INAR (1)) process, J. Time Ser. Anal. 8 (1987), pp. 261–275. [Google Scholar]
- 2.Bourguignon M., Rodrigues J., and Santos-Neto M., Extended Poisson INAR (1) processes with equidispersion, underdispersion and overdispersion, J. Appl. Stat. 46 (2019), pp. 101–118. [Google Scholar]
- 3.Bu R., McCabe B., and Hadri K., Maximum likelihood estimation of higher-order integer-valued autoregressive processes, J. Time Ser. Anal. 29 (2008), pp. 973–994. [Google Scholar]
- 4.Frey S. and Sandås P., The impact of iceberg orders in limit order books, in AFA 2009 San Francisco Meetings Paper, 2009.
- 5.Jung R.C. and Tremayne A., Useful models for time series of counts or simply wrong ones? Adv. Stat. Anal., 95 (2011), pp. 59–91. [Google Scholar]
- 6.Karlis D., EM algorithm for mixed poisson and other discrete distributions, ASTIN Bull. J. IAA 35 (2005), pp. 3–24. [Google Scholar]
- 7.Kirchner M., Hawkes and INAR(∞) processes, Stoch. Process. Appl. 126 (2016), pp. 2494–2525. [Google Scholar]
- 8.McKenzie E., Some simple models for discrete variate time series, J. Am. Water Resour. Assoc. 21 (1985), pp. 645–650. [Google Scholar]
- 9.McKenzie E., Autoregressive moving-average processes with negative-binomial and geometric marginal distributions, Adv. Appl. Probab. 18 (1986), pp. 679–705. [Google Scholar]
- 10.Ristić M.M., Bakouch H.S., and Nastić A.S., A new geometric first-order integer-valued autoregressive (NGINAR (1)) process, J. Stat. Plan. Inference 139 (2009), pp. 2218–2226. [Google Scholar]
- 11.Scotto M.G., Weiß C.H., and Gouveia S., Thinning-based models in the analysis of integer-valued time series: A review, Stat. Model. 15 (2015), pp. 590–618. [Google Scholar]
- 12.Weiß C.H., Thinning operations for modeling time series of counts a survey, Adv. Stat. Anal. 92 (2008), pp. 319–341. [Google Scholar]
- 13.Weiß C.H., A poisson INAR(1) model with serially dependent innovations, Metrika 78 (2015), pp. 829–851. [Google Scholar]
- 14.Zheng H., Basawa I.V., and Datta S., First-order random coefficient integer-valued autoregressive processes, J. Stat. Plan. Inference 137 (2007), pp. 212–229. [Google Scholar]



