Study the statistical properties of the output response of a system when the input is wide sense stationary

Random variable

A random variable is a mathematical formalization of a quantity or object which depends on random events.

 

Random process

A random process is a mathematical object that usually depends on a sequence defined as a sequence of random variables, where the index of the sequence has the interpretation of time. 

 

 

3. Definition of Stationary and Wide Sense Stationary Process

A stochastic process {…, Xt-1, Xt, Xt+1, Xt+2, …} consisting of random variables indexed by time index t is a time series.

The stochastic behavior of {Xt} is determined by specifying the probability density or mass functions (pdf’s).

p(xt1, ­ xt2, xt3, …, xtm)

for all finite collections of time indexes

{(t1, t2, …, tm), m < ∞}

i.e., all finite-dimensional distributions of {Xt}

 

A time series {Xt} is strictly stationary if

p(t1 + τ, t2 + τ, …, tm + τ) = p(t1, t2, …, tm),

τ, m, (t1, t2, …, tm).

Where p(t1 + τ, t2 + τ, …, tm + τ) represents the cumulative distribution function of the unconditional (i.e., with no reference to any particular starting value) joint distribution (parameters such as mean and variance) of {Xt} is said to be strictly stationary or strict-sense stationary if τ doesn’t affect the function p. So, p is not a function of time.

 

A time series {Xt} is called covariance stationary if

E(Xt) = μ

Var(Xt) = σ2x

Cov(Xt, Xt+τ) = γ(τ)

** (All constant over time t)

 

 

Wide Sense Stationary Process

A random process is called weak-sense stationary or wide-sense stationary (WSS) if its mean function and its correlation function do not change by shifts in time.

µx(t) = µx

Rxx(t1, t2) = Rxx(t1 + α, t2 + α)   for every α

                   = Rxx(t1 - t2, 0)

 

 

3. a. The output of a Wide Sense Stationary When the System is LTI

The output of filter is Y(t) = h(t) * X(t), where h(t) is the impulse response of a filter. Let the output mean be µy.

µy = E[Y(t)] = E[h(t) * X(t)]                                           

[‘E’ denotes the expected value]

µy = E[]

µy = ]

 

Now, if X(t) is WSS, a shift in time does not affect its mean, i.e.,

E[X(t)] = E[X(t - )] = µx

E[Y(t)] =  µx

            = µx

 

 = H(0)

 

µy = µx . H(0)

 

Since the H(0) will be some constant, we conclude that the output will also have the same nature as the input signal. 

So, the output of a wide sense stationary when the system is LTI is zero mean and WSS signal.

 

 

3. b. When the system is Linear and the input is Gaussian

An important property of the Gaussian random process is that their Probability density function is completely determined by their mean and covariance, i.e.,

fx(X) =  exp()

Where  is mean

And σ is standard deviation

 

The Fourier analysis states that any Linear system in the frequency spectrum can be represented as the sum of sinusoids, i.e.,

H(ω) = ω0)

Now, if the input X(t) is Gaussian, then the output y(t) in the frequency spectrum will be :

Y(ω) = X(ω).H(ω)

 

For a sinusoid input, the output will be varied in Magnitude and phase, but nature will remain the same according to the input.

 

So, the output will be Gaussian.

 

The output phase and magnitude will be varied, but nature will remain the same as that of the input. So, if the input is Gaussian, the output will also be Gaussian.

 

 

3. c. Time Series Analysis - AR, MA, and ARMA Models

 

Wold Representation Theorem

Any zero-mean covariance stationary time series {Xt}can be decomposed as

Xt = Vt  +  St

Where { Vt } is a linearly deterministic process, i.e., a linear combination of past values of  Vt with constant coefficients. It is analogical to autoregressive (AR) model.

St = it-i  is an infinite moving average process of error terms, where

θ0 = 1

i )2  < 

t is linearly unpredictable white noise, i.e.,

E(Ꜫt) = 0,  E(Ꜫt2) = σ2,  E(Ꜫt s) = 0  t,  s ≠ t

And t is uncorrelated with Vt

E(Ꜫt ­Vs) = 0  t, s

 

The output of a linear system characterized by a rational system function of the form

    … (i)

The corresponding difference equation is

…. (ii)

 

Where w[n] is the input sequence to the system and the observation data, x[n], represents the output sequence.

In power spectrum estimation, the input sequence is not observable. However, if the observed data are characterized as a stationary random process, then the input sequence is also assumed to be a stationary random process. In such a case the power density spectrum of the data is

   … (iii)

Where Γww(f) is the power density spectrum of the input sequence and H(f) is the frequency response of the model.

Since our objective is to estimate the power density spectrum Γxx(f), it is convenient to assume that the input sequence w[n] is a zero mean white noise sequence with autocorrelation.

Γww(m) = σ2w δ(m)    … (iv)

 

 

Where σ2w  is variance (i.e., σ2w = |w(n)|2). Then the power density spectrum of the observed data is simply

   … (v)

In the model-based approach the spectrum estimation procedure consists of two steps. Given data sequence x[n], 0 ≤ n ≤ N-1, we estimate the parameters {ak} and {bk} of the model. Then from this estimates, we compute the power spectrum estimate according to equation (v)

 

Random process x[n] generated by pole-zero model as shown in equation (i) is called an autoregressive moving average (ARMA) process of order (p,q) and it is usually denoted as ARMA (p,q). If q = 0 and b0 = 1, the resulting system model has a system function H(z) = 1 / A(z) and its output x[n] is called an autoregressive (AR) process of order p. This is denoted as AR(p). The third possible model is obtained by setting obtained by setting A(z) = 1, so that H(z) = B(z). Its output x[n] is called a moving average (MA) process of order q and denoted as MA(q).

Of these three linear models the AR model is by far the most widely used. The reasons are twofold. First, the AR model is suitable for representing spectra with narrow peaks (resonances). Second, the AR model results in very simple linear equations for the AR parameters. On the other hand, the MA model, as a general rule, requires many more coefficients to represent a narrow spectrum. Consequently, it is rarely used by itself as a model for spectrum estimation. By combining poles and zeros, the ARMA model provides a more efficient representation from the viewpoint of the number of model parameters, of the spectrum of a random process. The decomposition theorem due to Wold representation asserts that any ARMA or MA process can be represented uniquely by an AR model of possibly infinite order. In view of this theorem, the issue of model selection reduces to selecting the model that requires the smallest number of model parameters that are easy to compute. Usually the choice in practice is the AR model.

 

 

 

Autoregressive (AR) Model

The notation AR(p) indicates an autoregressive model of order p. The AR(p) model is defined as

Xt = i Xt-i  + Ꜫt

Where φ1, …, φp are parameters of the model, and Ꜫt is white noise. This can be equivalently written using the backshift operator B as

Xt = i Bi Xt  + Ꜫt

 

so that, moving the summation term to the left side and using polynomial notation, we have

ϕ[B] Xt  = Ꜫt

 

An autoregressive model is simply a linear regression of the current value of the series against one or more prior values of the series. The value of p is called the order of the AR model. AR models can be analyzed with one of various methods; including standard linear least squares techniques. They also have a straightforward interpretation.

 

An autoregressive model can thus be viewed as the output of an all-pole infinite impulse response filter whose input is white noise.

Some parameter constraints are necessary for the model to remain weak-sense stationary. For example, processes in the AR(1) model with |φ1| ≥ 1 are not stationary. More generally, for an AR(p) model to be weak-sense stationary, the roots of the polynomial Φ(z) : = 1 - i zi   must lie outside the unit circle i.e., each (complex) root zi must satisfy  | zi | > 1

 

 

Moving-average (MA) model

The notation MA(q) refers to the moving average model of order q:

Xt = µ + Ꜫt + θ1t-1 + … + θqt-q   =  µ + it-i + Ꜫt

 

Where µ is the mean of the series, the θ1, θ2, …, θq are parameters of the model and the Ꜫt, Ꜫt-1, …, Ꜫt-q  are white noise error terms. The value of q is called the order of the MA model. This can be equivalently written in terms of the backshift operator B as

Xt = µ + (1 + θ1B + … + θqBq)Ꜫt   

 

Thus a moving-average model is conceptually a linear regression of the current value of series against current and previous (observed) white noise error terms or random shocks. The random shocks at each point are assumed to be mutually independent and to come from the same distribution, typically a normal distribution, with location at zero and constant scale.

 

 

Autoregressive-moving-average (ARMA) Model

In the statistical analysis of time series, autoregressive-moving-average (ARMA) models provide a parsimonious description of a (weakly) stationary stochastic process in terms of two polynomials, one for autoregression (AR) and the second for moving-average (MA).

For a given time series of data Xt, the ARMA model is a tool for understanding and, perhaps, predicting future values in this series. The AR part involves regressing the variable on its own lagged (i.e., past) values. The MA part involves modeling the error term as a linear combination of error terms occurring contemporaneously and at various times in the past.

The model is usually referred to as the ARMA(p,q) model where p is the order of the AR part and q is the order of the MA part.

 

The notation ARMA(p, q) refers to the model with p autoregressive terms and q moving-average terms. This model contains the AR(p) and MA(q) models,

Xt = Ꜫt  +  i Xt-i   +  it-i

 

This method was useful for low-order polynomials (of degree three or less). The ARMA model is essentially an infinite impulse response filter applied to white noise.

 

 

Specifications in terms of lag operator

The model can be specified in terms of the lag operator L. In these terms then the AR(p) model is given by

t  =  (i Li ) Xt = φ(L) Xt

 

Where φ represents the polynomial

φ(L) = i Li

 

The MA(q) model is given by

Xt  =  ( 1  +   i Li  ) Ꜫt  =  θ(L) Ꜫt,

 

Where θ represents the polynomial

θ(L) = 1  +  i Li  

 

Finally, the combined ARMA(p,q) model is given by

 (i Li ) Xt   =   ( 1  +   i Li  ) Ꜫt

 

Or, more concisely,

φ(L)Xt   =   θ(L) Ꜫt

 

or,

 Xt  =  t

 

 

Applications

The time-series models are used to analyze and predict the data. A linear time series is modeled by linear difference equations involving the time series and the white noise or the innovation process. Such ARMA(p, q) models can be analyzed using the linear system theory.

Moreover, it is used for weather forecasting, climate forecasting, economic forecasting, healthcare forecasting engineering forecasting, finance forecasting, retail forecasting, business forecasting, environmental studies forecasting, social studies forecasting, and more.