# Time series with infinite-order partial copula dependence

• Martin Bladt and Alexander J. McNeil
From the journal Dependence Modeling

## Abstract

Stationary and ergodic time series can be constructed using an s-vine decomposition based on sets of bivariate copula functions. The extension of such processes to infinite copula sequences is considered and shown to yield a rich class of models that generalizes Gaussian ARMA and ARFIMA processes to allow both non-Gaussian marginal behaviour and a non-Gaussian description of the serial partial dependence structure. Extensions of classical causal and invertible representations of linear processes to general s-vine processes are proposed and investigated. A practical and parsimonious method for parameterizing s-vine processes using the Kendall partial autocorrelation function is developed. The potential of the resulting models to give improved statistical fits in many applications is indicated with an example using macroeconomic data.

AMS Subject Classification: 62M10; 62M05; 62H05; 60G10; 60G15; 60G22

## 1 Introduction

The principal aim of this article is to show that the s-vine (or stationary d-vine) decomposition of a joint density provides a very natural vehicle for generalizing the class of stationary Gaussian time series to permit both non-Gaussian marginal behaviour and non-linear and non-Gaussian serial dependence behaviour. In particular, this approach provides a route to defining a rich class of tractable non-Gaussian ARMA and ARFIMA processes; the resulting models have the potential to offer improved statistical fits in any application where classical ARMA models or their long-memory ARFIMA extensions are used.

Vine models of dependence have been developed in a series of publications [1,6,7, 8,24,25,28,42]. There are a number of different configurations for vines, but the most suitable one for longitudinal data applications is the d-vine, which is able to describe the strict stationarity of a random vector under some additional translation-invariance restrictions on the vine structure. A recent paper by Nagler et al. [36] investigated the vine structures that can be used to construct stationary multivariate time series. The results of Nagler et al. imply that, for univariate applications, the d-vine is in fact the only structure for which translation-invariance restrictions are sufficient to guarantee stationarity; we follow them in referring to these restricted d-vines as stationary vines, or s-vines.

Vine models are best understood as copula models of dependence, and there is now a large literature on copula models for time series. While the main focus of much of this literature has been on cross-sectional dependence between multiple time series, there is also a growing literature on modelling serial dependence within single series and lagged dependence across series. First-order Markov copula models [3,12,15,17] are simple examples of s-vine processes. A number of authors have written on higher-order Markov extensions for univariate series or multivariate series [4,10,22,30,36,43]. There is also literature showing how these models may be adapted to the particular requirements of time series showing stochastic volatility, including the mixture-copula approach of Loaiza-Maya et al. [30] and the v-transform approach of McNeil and Bladt [9,33].

This article makes the following novel contributions to the development of time series models based on vine copulas. First, we suggest how s-vine models may be generalized to infinite order, and we propose accompanying generalizations of the classical concepts of causality and invertibility for linear processes that may be applied to s-vine processes. Second, we provide additional insight into the issues of stability and ergodicity for s-vine processes, and we show how finite or infinite copula sequences may be used to develop non-linear filters of independent noise that generalize linear filters. Finally, we propose a practical and parsimonious approach to building s-vine processes in which copula sequences are parameterized by a function that we call the Kendall partial autocorrelation function; the latter may be borrowed from other well-known processes, such as Gaussian ARMA or ARFIMA processes, thus yielding natural non-Gaussian analogues of these models.

We believe that our approach may serve as a useful framework to faciliate further study in the field. Several interesting theoretical questions remain, particularly relating to necessary and sufficient conditions for the stability of models based on infinite copula sequences, as well as the interplay of copula sequences and long memory. However, on the practical side, the models are already eminently usable; methods exist for estimation and random number generation, and we suggest some new ideas for model validation using residuals. An example shows the benefits that may arise from using these models.

This article is structured as follows. Section 2 sets out notation and basic concepts and makes the connection between s-vine copulas and s-vine processes; key objects in the development of processes are sequences of functions that we refer to as Rosenblatt functions. In Section 3, we show that finite-order s-vine processes are Markov chains belonging to the particular sub-category of non-linear state-space models. Section 4 explains why Gaussian processes form a sub-class of s-vine processes and shows how the classical theory for linear processes may be reinterpreted as a theory of the behaviour of Rosenblatt functions. Section 5 uses the Gaussian analogy to suggest requirements for stable, infinite-order, non-Gaussian s-vine processes; a practical approach to model building is developed and illustrated with an application to macroeconomic data. Section 6 concludes. Proofs can be found in Appendix A, while additional material on the Markov chain analysis of finite-order processes is collected in Appendix B.

## 2 S-vine processes

### 2.1 S-vine copulas

If a random vector ( X 1 , , X n ) admits a joint density f ( x 1 , , x n ) then the latter may be decomposed as a d-vine. Writing f X i for the marginal density of X i , the decomposition is

(1) f ( x 1 , , x n ) = i = 1 n f X i ( x i ) k = 1 n 1 j = k + 1 n c j k , j S j k , j ( F j k S j k , j ( x j k ) , F j S j k , j ( x j ) ) ,

where S j k , j = { j k + 1 , , j 1 } is the set of indices of the variables which lie between X j k and X j , c j k , j S j k , j is the density of the bivariate copula C j k , j S j k , j of the joint distribution function (df) of X j k and X j conditional on the intermediate variables X j k + 1 , , X j 1 , and

(2) F i S j k , j ( x ) = P ( X i x X j k + 1 = x j k + 1 , , X j 1 = x j 1 ) , i { j k , j }

denotes the conditional df of variable i conditional on these variables; note that S j 1 , j = and so the conditioning set is dropped in this case. The decomposition in Eq. (1) implies a decomposition of the density c ( u 1 , , u n ) of the unique copula of ( X 1 , , X n ) , which is given implicitly by

(3) c ( F 1 ( x 1 ) , , F n ( x n ) ) = k = 1 n 1 j = k + 1 n c j k , j S j k , j ( F j k S j k , j ( x j k ) , F j S j k , j ( x j ) ) .

In practical applications, interest centres on models that admit the simplified d-vine decomposition in which the copula densities c j k , j S j k , j do not depend on the values of variables in the conditioning set S j k , j and we can simply write c j k , j . Any set of copula densities { c j k , j : 1 k n 1 , k + 1 j n } and any set of marginal densities f X i may be used in the simplified version of (1) to create a valid n -dimensional joint density. A number of papers have examined the limitations imposed by working with simplified vine copula models [20,35,44,45]. In Mroz et al. [35], it is shown that the class of simplified vines is not dense in the space of copulas for a number of metrics including the one induced by total variation distance. These results may be interpreted as showing that there exist multivariate distributions that are difficult to approximate with simplified d-vines. However, the simplified d-vine construction still greatly enlarges the class of tractable densities for time series applications.

We are interested in strictly stationary stochastic processes whose higher-dimensional marginal distributions are simplified d-vines. As well as forcing f X 1 = = f X n , this requirement imposes translation-invariance conditions on the copula densities c j k , j and conditional dfs F S j k , j appearing in the simplified form of Eq. (1). It must be the case that c j k , j is the same for all j { k + 1 , , n } , and so each pair copula density in the model can be associated with a lag k and we can write c k c j k , j , where c k is the density of some bivariate copula C k . The conditional dfs can be represented by two sets of functions R k ( 1 ) : ( 0 , 1 ) k × ( 0 , 1 ) ( 0 , 1 ) and R k ( 2 ) : ( 0 , 1 ) k × ( 0 , 1 ) ( 0 , 1 ) , which are defined in a recursive, interlacing fashion by R 1 ( 1 ) ( u , x ) = h 1 ( 1 ) ( u , x ) , R 1 ( 2 ) ( u , x ) = h 1 ( 2 ) ( x , u ) , and, for k 2 ,

(4) R k ( 1 ) ( u , x ) = h k ( 1 ) ( R k 1 ( 2 ) ( u 1 , u 1 ) , R k 1 ( 1 ) ( u 1 , x ) ) , R k ( 2 ) ( u , x ) = h k ( 2 ) ( R k 1 ( 2 ) ( u k , x ) , R k 1 ( 1 ) ( u k , u k ) ) ,

where h k ( i ) ( u 1 , u 2 ) = u i C k ( u 1 , u 2 ) and u i indicates the vector u with the i th component removed.

By using this new notation, we obtain a simplified form of Eq. (1) in which the density of the copula c in Eq. (3) takes the form

(5) c ( n ) ( u 1 , , u n ) = k = 1 n 1 j = k + 1 n c k ( R k 1 ( 2 ) ( u [ j k + 1 , j 1 ] , u j k ) , R k 1 ( 1 ) ( u [ j k + 1 , j 1 ] , u j ) ) ,

where u [ j k + 1 , j 1 ] = ( u j k + 1 , , u j 1 ) . Note that, for simplicity of formulas, we abuse notation by including terms involving R 0 ( 1 ) and R 0 ( 2 ) ; these terms should be interpreted as R 0 ( 1 ) ( , u ) = R 0 ( 2 ) ( , u ) = u for all u . Following Nagler et al. [36], we refer to a model with copula density of the form Eq. (5) as a stationary d-vine or s-vine.

If a random vector ( U 1 , , U n ) follows the copula C ( n ) with density c ( n ) in Eq. (5), then for any k { 1 , , n 1 } and j { k + 1 , , n } , we have

(6) R k ( 1 ) ( u , x ) = P ( U j x U j k = u 1 , , U j 1 = u k ) , R k ( 2 ) ( u , x ) = P ( U j k x U j k + 1 = u 1 , , U j = u k ) ,

and we refer to the conditional distribution functions R k ( 1 ) and R k ( 2 ) as forward and backward Rosenblatt functions. Henceforth, we will often drop the superscript from the forward function and simply write R k = R k ( 1 ) to obtain less notationally cumbersome expressions. The conditional densities corresponding to the Rosenblatt functions may be derived from Eq. (5). Writing f k for the density of the forward Rosenblatt functions, we obtain f 1 ( u , x ) = c ( 2 ) ( u , x ) = c 1 ( u , x ) and, for k > 1

(7) f k ( u , x ) = c ( k + 1 ) ( u 1 , , u k , x ) c ( k ) ( u 1 , , u k ) = j = 1 k c j ( R j 1 ( 2 ) ( u [ k j + 2 , k ] , u k j + 1 ) , R j 1 ( u [ k j + 2 , k ] , x ) ) .

The following assumption will be in force throughout the remainder of the paper.

### Assumption 1

All copulas C k used in the construction of s-vine models belong to the class C of smooth functions with continuous partial derivatives of all orders. Moreover, their densities c k are strictly positive on ( 0 , 1 ) 2 .

This assumption applies to all the standard pair copulas that are used in vine copula models (e.g., Gauss, Clayton, Gumbel, Frank, Joe, and t), as well as non-exchangeable extensions [29] or mixtures of copulas [30]. It ensures, among other things, that for fixed u , the Rosenblatt functions are bijections on ( 0 , 1 ) with well-defined inverses. Let us write R k 1 ( u , z ) for the inverses of the Rosenblatt forward functions, satisfying R k 1 ( u , z ) = x if and only if R k ( u , x ) = z . Inverses can also be defined for the Rosenblatt backward functions but will not be explicitly needed

In the sequel, we refer to the copulas C k as partial copulas. They should be distinguished from the bivariate marginal copulas given by C ( k ) ( u , v ) = P ( U j k u , U j v ) for any j { k + 1 , , n } . The two copulas are related by the formula

(8) C ( k ) ( v 1 , v 2 ) = E ( P ( U j k v 1 , U j v 2 U j k + 1 , , U j 1 ) ) = E ( C k ( R k 1 ( 2 ) ( ( U j k + 1 , , U j 1 ) , v 1 ) , R k 1 ( ( U j k + 1 , , U j 1 ) , v 2 ) ) ) = 0 1 0 1 C k ( R k 1 ( 2 ) ( u , v 1 ) , R k 1 ( u , v 2 ) ) c ( k 1 ) ( u ) d u 1 d u k 1 .

### 2.2 S-vine processes

We use the following general definition for an s-vine process.

### Definition 1

(S-vine process) A strictly stationary time series ( X t ) t Z is an s-vine process if for every t Z and n 2 the n -dimensional marginal distribution of the vector ( X t , , X t + n 1 ) is absolutely continuous and admits a unique copula C ( n ) with a joint density c ( n ) of the form in Eq. (5). An s-vine process ( U t ) t Z is an s-vine copula process if its univariate marginal distribution is standard uniform.

Our aim is to construct processes that conform to this definition and investigate their properties and practical application. Since s-vine processes can be endowed with any continuous univariate marginal distribution f X , we will mostly investigate the properties of s-vine copula processes.

### 2.3 A note on reversibility

It is particularly common in applications of vine copulas to confine interest to standard exchangeable copulas C k . In this case, the resulting s-vine processes have the property of reversibility. For any u = ( u 1 , , u n ) ( 0 , 1 ) n , let us write u ¯ = ( u n , , u 1 ) for the reversed vector.

### Definition 2

An s-vine copula process is reversible if for any n 2 the higher dimensional marginal copulas satisfy C ( n ) ( u ) = C ( n ) ( u ¯ ) .

This is equivalent to saying that, for any t , s Z and any n > 2 , the set of consecutive variables ( U t + 1 , , U t + n ) from the process has the same distribution as the reversed vector ( U s + n , , U s + 1 ) . The process evolves forwards and backwards in a similar fashion, which may not be ideal for phenomena in which there is a clear temporal notion of causality; however, as soon as non-exchangeable copulas are included, the reversibility is broken. In summary, we have the following simple result.

### Proposition 1

If a copula sequence ( C k ) k N consists of exchangeable copulas then (i) the Rosenblatt forward and backward functions satisfy R k ( 2 ) ( u ¯ , x ) = R k ( u , x ) for all ( u , x ) ( 0 , 1 ) k × ( 0 , 1 ) and (ii) the resulting s-vine copula process is reversible.

## 3 S-vine processes of finite order

### 3.1 Markov construction

The first class of processes we consider are s-vine copula processes of finite order p which are constructed from a set of copulas { C 1 , , C p } using the Markov approach described by Joe ([27], p. 145). Starting from a series of iid uniform innovation variables ( Z k ) k N we can set U 1 = Z 1 and

(9) U k = R k 1 1 ( ( U 1 , , U k 1 ) , Z k ) , k 2 .

By using the inverses of the Rosenblatt forward functions we obtain, for any n , a random vector ( U 1 , , U n ) which forms a finite realization from an s-vine process ( U t ) t Z . The copula C ( n ) of ( U 1 , , U n ) has density c ( n ) in Eq. (5) but the copula densities c k appearing in this expression satisfy c k ( u , v ) = 1 for k > p and the s-vine is said to be truncated at order p . Moreover, since h k ( 1 ) ( u , v ) = v for k > p , it follows from Eq. (4) that R k ( u , x ) = R k 1 ( u 1 , x ) = = R p ( u [ k p + 1 , k ] , x ) and the updating Eq. (9) satisfies

(10) U k = R p 1 ( ( U k p , , U k 1 ) , Z k ) , k > p ,

showing the Markovian character of the finite-order process.

The recursive nature of the construction (Eq. (9)) means that there is an implied set of functions that we will label S k : ( 0 , 1 ) k × ( 0 , 1 ) ( 0 , 1 ) for k N such that

(11) U k = S k 1 ( ( Z 1 , , Z k 1 ) , Z k ) , k 2 .

The functions ( S k ) k N satisfy S 1 ( z 1 , x ) = R 1 1 ( z 1 , x ) and

(12) S k ( z , x ) = R k 1 ( ( z 1 , S 1 ( z 1 , z 2 ) , , S k 1 ( z [ 1 , k 1 ] , z k ) ) , x ) , k 2 .

The identity in Eq. (11) can be thought of as a causal representation of the process, while the complementary identity Z k = R k 1 ( ( U 1 , , U k 1 ) , U k ) implied by Eq. (9) can be thought of as an invertible representation. We refer to the functions ( S k ) k N as Rosenblatt inverse functions; they should be distinguished from the inverses of the Rosenblatt forward functions

### 3.2 Non-linear state space model

The s-vine process of order p can be viewed as a p -dimensional Markov chain with state space X = ( 0 , 1 ) p . It is standard to treat Markov chains as being indexed by the natural numbers. To that end, for t N , we introduce the vector-valued process U t = ( U t , , U t + p 1 ) , starting at U 1 = ( U 1 , , U p ) , defined by the updating equation U t = F ( U t 1 , Z t ) , where

(13) F : ( 0 , 1 ) p × ( 0 , 1 ) ( 0 , 1 ) p , F ( u , z ) = ( u 2 , , u p , R p 1 ( u , z ) ) .

The Markov chain described by Eq. (13) defines a non-linear state space (NSS) model conforming exactly to the assumptions imposed in Meyn and Tweedie ([34], Section 2.2.2): under Assumption 1, the updating function F is a smooth ( C ) function; the state space X = ( 0 , 1 ) p is an open subset of R p ; the uniform distribution of innovations ( Z t ) will be taken to be supported on the open set ( 0 , 1 ) .

Using standard arguments, the NSS model associated with Eq. (13) can be shown to be a ϕ -irreducible, aperiodic Harris recurrent Markov chain and to admit an invariant probability measure π , which is the measure implied by the density c ( p ) given by Eq. (5); we summarise the arguments in Appendix B. This in turn allows the ergodic theorem for Harris chains to be applied ([34], Theorem 13.3.3) to conclude that for any initial measure λ , the Markov transition kernel P ( x , ) satisfies

λ ( d x ) P n ( x , ) π ( ) 0 , n ,

where denotes the total variation norm. This is also sufficient for the strong law of large numbers (SLLN) to hold ([34], Theorem 17.0.1): for a function g : R p R , if we define S n ( g ) = k = 1 n g ( U k ) and π ( g ) = g ( u ) c ( p ) ( u ) d u , then lim n n 1 S n ( g ) = π ( g ) , almost surely, provided π ( g ) < .

Although the Markov models are ergodic, we caution that they can exhibit some very extreme behaviour, albeit for copula choices that we are unlikely to encounter in practice. Figure 1 shows a realisation of 10,000 simulated values from a process of order p = 3 , in which C 1 is a 180-degree rotated Clayton copula with parameter θ = 2 , C 2 is a Clayton copula with θ = 2 , and C 3 is a rotated Clayton copula with θ = 4 . Since the Clayton copula is well known to have lower tail dependence [25,27], this means that C 1 and C 3 have upper tail dependence and C 3 is more strongly dependent than C 1 and C 2 . This increasing pattern of partial dependence, coupled with the strong upper tail dependence of C 3 , leads to a period of over 1,500 successive values, which are all greater than 0.6. An observer of this process who plots a histogram of the values in this period would have difficulty believing that the marginal distribution is uniform.

Figure 1

Realisation of 10,000 simulated values from a process of order k = 3 in which C 1 is a 18 0 rotated Clayton copula with parameter θ = 2 , C 2 is a Clayton copula with θ = 2 and C 2 is a rotated Clayton copula with θ = 4 .

This phenomenon is connected to rates of mixing behaviour and ergodic convergence for Markov processes. There is some literature for the case p = 1 in which these rates are shown to vary with the choice of copula and, in particular, its behaviour in joint tail regions [3,5,12,13,31]. For some results relevant to the case, where p > 1 , see Rémillard et al. [39].

## 4 Gaussian processes

Gaussian processes are processes whose finite-dimensional marginal distributions are multivariate Gaussian. We will identify the term Gaussian processes with non-singular Gaussian processes throughout; i.e., we assume that the finite-dimensional marginal distributions of Gaussian processes have invertible covariance matrices and admit joint densities. Such processes represent a subclass of the s-vine processes.

## Proposition 2

1. Every stationary Gaussian process is an s-vine process.

2. Every s-vine process in which the pair copulas of the sequence ( C k ) k N are Gaussian and the marginal distribution F X is Gaussian, is a Gaussian process.

### 4.1 S-vine representations of Gaussian processes

The first implication of Proposition 2 is that every Gaussian process has a unique s-vine-copula representation. This insight offers methods for constructing or simulating such processes as generic s-vine processes using Eq. (9) and estimating them using a likelihood based on Eq. (5).

Let ( X t ) t N be a stationary Gaussian process with mean μ X , variance σ X 2 , and autocorrelation function (acf) ( ρ k ) k N ; these three quantities uniquely determine a Gaussian process. We assume the following:

### Assumption 2

The acf ( ρ k ) k N satisfies ρ k 0 as k .

It is well known that this is a necessary and sufficient condition for a Gaussian process ( X t ) to be a mixing process and therefore ergodic [14,32].

The acf uniquely determines the partial autocorrelation function (pacf) ( α k ) k N through a one-to-one transformation [2,38]. Since the partial autocorrelation of a Gaussian process is the correlation of the conditional distribution of ( X t k , X t ) given the intervening variables, the pair copulas in the s-vine copula representation are given by C k = C α k Ga .

For k N let ρ k = ( ρ 1 , , ρ k ) and let P k denote the correlation matrix of ( X 1 , , X k ) . Clearly, P 1 = 1 and, for k > 1 , P k is a symmetric Toeplitz matrix whose diagonals are filled by the first k 1 elements of ρ k ; moreover, P k is non-singular for all k under Assumption 2 ([11], Proposition 4). The one-to-one series of recursive transformations relating ( α k ) k N to ( ρ k ) k N is α 1 = ρ 1 , and, for k > 1 ,

(14) α k = ρ k ρ k 1 P k 1 1 ρ ¯ k 1 1 ρ k 1 P k 1 1 ρ k 1 , ρ k = α k ( 1 ρ k 1 P k 1 1 ρ k 1 ) + ρ k 1 P k 1 1 ρ ¯ k 1 ;

see, for example, Joe [26] or the Durbin–Levinson Algorithm ([11], Proposition 5.2.1).

### Remark 1

Note that the restriction to non-singular Gaussian processes ensures that ρ k < 1 and α k < 1 , for all k N , and this is henceforth always assumed.

We review three examples of well-known Gaussian processes from the point of view of s-vine processes.

### Example 1

(Gaussian ARMA models) Any causal Gaussian ARMA( p , q ) model may be represented as an s-vine process, and full maximum likelihood estimation can be carried out using a joint density based on Eq. (5). If ϕ = ( ϕ 1 , , ψ p ) and ψ = ( ψ 1 , , ψ q ) denote the AR and MA parameters and ρ k ( ϕ , ψ ) the acf, then we can use the transformation in Eq. (14) to parameterize Eq. (5) in terms of ϕ and ψ using Gaussian pair copulas C k = C α k ( ϕ , ψ ) Ga . In practice, this approach is more of theoretical interest since standard estimation methods are generally much faster.

### Example 2

(Fractional Gaussian noise [FGN]) This process has acf given by

ρ k ( H ) = 1 2 ( ( k + 1 ) 2 H + ( k 1 ) 2 H 2 k 2 H ) , 0 < H < 1 ,

where H is the Hurst exponent [41]. Thus, the transformation Eq. (14) may be used to parameterize Eq. (5) in terms of H using Gaussian pair copulas C k = C α k ( H ) Ga and the FGN model may be fitted to data as an s-vine process and H may be estimated.

### Example 3

(Gaussian ARFIMA models) The ARFIMA( p , d , q ) model with 1 / 2 < d < 1 / 2 can be handled in a similar way to the ARMA( p , q ) model, of which it is a generalization. In the case where p = q = 0 , it has been shown [21] that

(15) α k = d k d , k N ;

see also Brockwell and Davis ([11], Theorem 13.2.1). The simple closed-form expression for the pacf means that the ARFIMA( 0 , d , 0 ) model is even more convenient to treat as an s-vine than FGN; the two models are in fact very similar in behaviour although not identical. It is interesting to note that the pacf is not summable and similar behaviour holds for some other ARFIMA processes. For example, for p , q N { 0 } and 0 < d < 1 / 2 , the pacf satisfies α k d / k as k [23].

### 4.2 New Gaussian processes from s-vines

A further implication of Proposition 2 is that it shows how we can create and estimate some new stationary and ergodic Gaussian processes without setting them up in the classical way using recurrence equations, lag operators, and Gaussian innovations. Instead we choose sequences of Gaussian pair copulas ( C k ) parameterized by sequences of partial correlations ( α k ) .

As in the previous section, we can begin with a parametric form for the acf ρ k ( θ ) such that ρ k ( θ ) 0 as k and build the model using pair copulas parameterized by the parameters θ of the implied pacf α k ( θ ) . Alternatively we can choose a parametric form for the pacf α k ( θ ) directly.

Any finite set of values { α 1 , , α p } yields an AR(p) model, which is a special case of the finite-order s-vine models of Section 3. However, infinite-order processes that satisfy Assumption 2 are more delicate to specify. A necessary condition is that the sequence ( α k ) satisfies α k 0 as k 0 , but this is not sufficient. To see this, note that if α k = ( k + 1 ) 1 , the relationship (14) implies that ρ k = 0.5 for all k , which violates Assumption 2. A sufficient condition follows from a result of Debowski [16], although, in view of Example 3, it is not a necessary condition:

### Assumption 3

The partial acf ( α k ) k N satisfies k = 1 α k < .

Debowski [16] showed that, if Assumption 3 holds, then the equality

(16) 1 + 2 k = 1 ρ k = k = 1 1 + α k 1 α k

also holds. The rhs of Eq. (16) is a convergent product since absolute summability ensures that the sums k = 1 ln ( 1 ± α k ) converge. This implies the convergence of k = 1 ρ k , which implies ρ k 0 , which in turn implies that Assumption 2 also holds, as we require.

Assumption 3 still allows some quite pathological processes, as noted by Debowski [16]. For example, even for a finite-order AR( p ) process with α k a > 0 for k { 1 , , p } and α k = 0 for k > p , it follows that k = 1 ρ k 0.5 ( ( ( 1 + a ) / ( 1 a ) ) p 1 ) , and this grows exponentially with p leading to an exceptionally slow decay of the acf.

### 4.3 Rosenblatt functions for Gaussian processes

For Gaussian processes, the Rosenblatt functions and inverse Rosenblatt functions take relatively tractable forms.

### Proposition 3

Let ( C k ) k N be a sequence of Gaussian pair copulas with parameters ( α k ) k N and assume that Assumption 2holds. The forward Rosenblatt functions are given by

(17) R k ( u , x ) = Φ Φ 1 ( x ) j = 1 k ϕ j ( k ) Φ 1 ( u k + 1 j ) σ k ,

where σ k 2 = j = 1 i ( 1 α j 2 ) and the coefficients ϕ j ( k ) are given recursively by

(18) ϕ j ( k ) = ϕ j ( k 1 ) α k ϕ k j ( k 1 ) , j { 1 , , k 1 } , α k , j = k .

The inverse Rosenblatt functions are given by

(19) S k ( z , x ) = Φ σ k Φ 1 ( x ) + j = 1 k ψ j ( k ) Φ 1 ( z k + 1 j ) ,

where the coefficients ψ j ( k ) are given recursively by

(20) ψ j ( k ) = i = 1 j ϕ i ( k ) ψ j i ( k i ) , j { 1 , , k } ,

where ψ 0 ( k ) = σ k for k 1 and ψ 0 ( 0 ) = 1 .

We can analyse the behaviour of the Rosenblatt and inverse Rosenblatt functions as k in a number of different cases.

#### 4.3.1 Gaussian processes of finite order

In the case of a Gaussian s-vine process of finite-order p , we have, for k > p , that α k = 0 , σ k = σ p and ϕ j ( k ) = ϕ j ( p ) . If ( U k ) k N is constructed from ( Z k ) k N using the algorithm described by Eq. (9), and if we make the substitutions X k = Φ 1 ( U k ) and ε k = Φ 1 ( Z k ) as in the proof of Proposition 3, then it follows from Eq. (17) that X k = j = 1 p ϕ j ( p ) X k j + σ p ε k for k > p , which is the classical recurrence equation that defines a Gaussian AR( p ) process; from Eqs. (11) and (19), we also have that X k = j = 1 k 1 ψ j ( k 1 ) ε k j + σ p ε k for k > p . These two representations can be written in invertible and causal forms as follows:

(21) ε k = j = 0 p ϕ ˜ j ( p ) X k j and X k = j = 0 k 1 ψ j ( k 1 ) ε k j , k > p ,

where ϕ ˜ 0 ( p ) = 1 / σ p , ϕ ˜ j ( p ) = ϕ j ( p ) / σ p for j > 1 and ψ 0 ( k 1 ) = σ p .

The first series in Eq. (21) is clearly a finite series, while the classical theory is concerned with conditions on the AR coefficients ϕ ˜ j ( p ) that allow us to pass to an infinite-order moving-average representation as k in the second series. In fact, by setting up our Gaussian models using partial autocorrelations, causality in the classical sense is guaranteed; this follows as a special case of Theorem 1.

#### 4.3.2 Gaussian processes with absolutely summable partial autocorrelations

We next consider a more general case where the process may be of infinite order, but Assumption 3 holds. To consider infinite-order models, we now consider a process ( U t ) t Z defined on the integers. The result that follows is effectively a restating of a result by Debowski [16] in the particular context of Gaussian s-vine copula processes.

#### Theorem 1

Let ( U t ) t Z be a Gaussian s-vine copula process for which the parameters ( α k ) k N of the Gaussian pair copula sequence ( C k ) k N satisfy Assumption 3. Then, for all t , we have the almost sure limiting representations

(22) U t = lim k S k ( ( Z t k , , Z t 1 ) , Z t )

(23) Z t = lim k R k ( ( U t k , , U t 1 ) , U t )

for an iid uniform innovation process ( Z t ) t Z .

#### 4.3.3 Long-memory ARFIMA processes

As noted earlier, the pacf of an ARFIMA( p , d , q ) model with 0 < d < 0.5 is not absolutely summable [23], and so Theorem 1 does not apply in this case. Nevertheless, Brockwell and Davis ([11], Section 13.2) show that the Gaussian process has a casual representation of the form X t = j = 0 ψ j ε t j , where convergence is now in mean square and the coefficients are square summable, i.e., j = 0 ψ j 2 < . Since convergence in mean square implies convergence in probability, the continuous mapping theorem implies that a representation of the form U t = lim k S k ( ( Z t k , , Z t 1 ) , Z t ) at least holds under convergence in probability.

#### 4.3.4 A non-causal and non-invertible case

If α k = 1 / ( k + 1 ) for all k , then ρ k = 0.5 , and both Assumptions 2 and 3 are violated. It can be verified (for example by induction) that the recursive formulas (18) and (20) imply that ϕ j ( k ) = 1 / ( k + 1 ) and ψ j ( k ) = σ k j / ( k + 2 j ) for j 1 (recall that ψ 0 ( k ) = σ k ). These coefficient sequences are unusual; the coefficients ϕ j ( k ) of the Rosenblatt function in Eq. (17) place equal weight on all past values X k + 1 j = Φ 1 ( U k + 1 j ) , while the coefficients ψ j ( k ) of the inverse Rosenblatt function on the innovations in Eq. (19) place weight ψ k ( k ) = 1 / 2 on the first value ε 1 = Φ 1 ( Z 1 ) and decreasing weights on more recent values ε j , j > 1 .

As k , we do have σ k 2 = j = 1 k ( 1 1 / ( k + 1 ) 2 ) 1 / 2 , but, for fixed j 1 , the terms ψ j ( k ) and ψ j ( k ) both converge to the trivial limiting value 0. In particular, we do not obtain a convergent limiting representation of the form in Eq. 22.

## 5 General s-vine processes

We now consider infinite-order s-vine copula processes constructed from general sequences ( C k ) k N of pair copulas.

### 5.1 Causality and invertibility

The key consideration for the stability of an infinite-order process is whether it admits a convergent causal representation. A process ( U t ) t Z with such a representation is a convergent non-linear filter of independent noise. It will have the property that U t and U t k are independent in the limit as k , implying mixing behaviour and ergodicity. We suggest the following definition of the causality and invertibility properties for a general s-vine process.

### Definition 3

Let ( C k ) k N be a sequence of pair copulas and let ( R k ) k N and ( S k ) k N be the corresponding Rosenblatt forward functions and Rosenblatt inverse functions defined by Eqs. (4) and (12). An s-vine copula process ( U t ) t Z associated with the sequence ( C k ) k N is strongly causal if there exists a process of iid uniform random variables ( Z t ) t Z such that Eq. (22) holds almost surely for all t , and it is strongly invertible if representation (Eq. (23)) holds almost surely for all t . If convergence in Eqs. (22) and (23) only holds in probability, the process is weakly causal or weakly invertible.

We know that Gaussian ARMA processes defined as s-vine processes are always strongly causal (and invertible) and that the long-memory ARFIMA( p , d , q ) process with 0 < d < 0.5 is weakly causal. When we consider sequences of Rosenblatt functions for sequences of non-Gaussian pair copulas, proving causality appears to be more challenging mathematically, since it is no longer a question of analysing the convergence of series. In the next section, we use simulations to conjecture that causality holds for a class of processes defined via the Kendall correlations of the copula sequence.

In a finite-order process, the copula sequence for any lag k greater than the order p consists of independence copulas; it seems intuitively clear that, to obtain an infinite-order process with a convergent causal representation, the partial copula sequence ( C k ) k N should converge to the independence copula C as k . However, in view of Example 4.3.4, this is not a sufficient condition and the speed of convergence of the copula sequence is also important. Ideally, we require conditions on the speed of convergence C k C so that the marginal copula C ( k ) in Eq. (8) also tends to C ; in that case, the variables U t and U t k are asymptotically independent as k and mixing behaviour follows.

### 5.2 A practical approach to non-Gaussian s-vines

Suppose we take a sequence of pair copulas ( C k ) k N from some parametric family and parameterize them in such a way that (i) the copulas converge uniformly to the independence copula as k and (ii) the level of dependence of each copula C k is identical to that of a Gaussian pair copula sequence that gives rise to an ergodic Gaussian process. The intuition here is that by sticking close to the pattern of decay of dependence in a well-behaved Gaussian process, we might hope to construct a stable causal process that is both mixing and ergodic.

A natural way of making “level of dependence” concrete is to consider the Kendall rank correlation function of the copula sequence, defined in the following way.

### Definition 4

The Kendall partial autocorrelation function (kpacf) ( τ k ) k N associated with a copula sequence ( C k ) k N is given by τ k = τ ( C k ) , for k N , where τ ( C ) denotes the Kendall’s tau coefficient for a copula C .

For a Gaussian copula sequence with C k = C α k Ga , we have

(24) τ k = 2 π arcsin ( α k ) .

As in Section 4.2, suppose that ( α k ( θ ) ) k N is the pacf of a stationary and ergodic model Gaussian process parametrized by the parameters θ , such as an ARMA or ARFIMA model; this implies a parametric form for the kpacf ( τ k ( θ ) ) k N . The idea is to choose a sequence of non-Gaussian pair copulas that shares this kpacf.

A practical problem that may arise is that τ k = τ k ( θ ) can take any value in ( 1 , 1 ) in practice; only certain copula families, such as Gauss and Frank, are said to be comprehensive and yield any value for τ k . If we wish to use, for example, a sequence of Gumbel copulas to build our model, then we need to find a solution for negative values of Kendall’s tau. One possibility is to allow 90 or 270 degree rotations of the copula at negative values of τ k and another is to substitute a comprehensive copula at any position k in the sequence such that τ k is negative.

### Remark 2

Note that the assumption that the pair copulas C k converge to the independence copula has implications for using t copulas C ν , α t in this approach. The terms of the copula sequence C k = C ν k , α k t would have to satisfy ν k and α k 0 as k ; the sequence given by C k = C ν , α k t for fixed ν does not converge to the independence copula as α k 0 . While the sequence ( α k ) k N can be connected to the kpacf by the same formula (24), the sequence ( ν k ) k N is not fixed by the kpacf. It is simpler in this approach to work with copula families with a single parameter so that there is a one-to-one relationship between Kendall’s tau and the copula parameter.

To compare the speed of convergence of the copula filter for different copula sequences sharing the same kpacf, we conduct some simulation experiments. For fixed n and for a fixed realization z 1 , , z n of independent uniform noise we plot the points ( k , S k ( z [ n k , n 1 ] , z n ) ) for k { 1 , , n 1 } . We expect the points to converge to a fixed value as k n 1 , provided we take a sufficiently large value of n . When the copula sequence consists of Clayton copulas we will refer to the model as a Clayton copula filter; similarly, Gumbel copulas yield a Gumbel copula filter; and so on. The following examples suggest that there are some differences in the convergence rates of the copula filters. This appears to relate to the tail dependence characteristics of the copulas [25,27]. We recall that the Gumbel and Joe copulas are upper tail dependent, while the Clayton copula is lower tail dependent; the Gauss and Frank copulas are tail independent. The filters based on sequences of tail-dependent copulas generally show slower convergence.

### Example 4

(Non-Gaussian ARMA(1,1) models). In this example, we consider s-vine copula processes sharing the kpacf of the ARMA(1,1) model with autoregressive parameter 0.95 and moving-average parameter -0.85. Fixing n = 201 , we obtain Figure 2. Convergence appears to be fastest for the Gaussian and Frank copula filters and slowest for the Clayton filter, followed by the Joe filter; the Gumbel filter is an intermediate case. We can also discern a tendency for jumps in the value of S k ( z [ n k , n 1 ] , z n ) to be upward for the upper tail-dependent Gumbel and Joe copulas and downward for the lower tail-dependent Clayton copula.

Figure 2

Plots of ( k , S k ( z [ n k , n 1 ] , z n ) ) for k { 1 , , n 1 } for the copula filters of ARMA(1,1) models; see Example 4. Horizontal lines show ultimate values S n 1 ( z [ 1 , n 1 ] , z n ) .

### Example 5

(Non-Gaussian ARFIMA(1, d , 1) models) In this example, we consider s-vine copula processes sharing the kpacf of the ARFIMA(1, d , 1) model with autoregressive parameter 0.95, moving-average parameter 0.85 and fractional differencing parameter d = 0.02 . The latter implies that the pacf of the Gaussian process satisfies α k 0.02 / k as k [23]. The lack of absolute summability means that the Gaussian copula process does not satisfy the conditions of Theorem 1. It is an unresolved question whether any of these processes is causal. Fixing n = 701 , we obtain Figure 3. For the realized series of innovations used in the picture, convergence appears to take place, but it is extremely slow. The tail-dependent Clayton and Joe copulas appear to take longest to settle down.

Figure 3

Plots of ( k , S k ( z [ n k , n 1 ] , z n ) ) for k { 1 , , n 1 } for the copula filters of ARFIMA(1, d , 1) models; see Example 5. Horizontal lines show ultimate values S n 1 ( z [ 1 , n 1 ] , z n ) .

An obvious practical solution that circumvents the issue of whether the infinite-order process has a convergent causal representation is to truncate the copula sequence ( C k ) k N so that C k = C for k > p for some relatively large but fixed value p . This places us back in the setting of ergodic Markov chains but, by parameterizing models through the kpacf, we preserve the advantages of parsimony.

### 5.3 An example with real data

For this example, we have used data on the US CPI (consumer price index) taken from the OECD webpage. We analyse the log-differenced time series of quarterly CPI values from the first quarter of 1960 to the 4th quarter of 2020, which can be interpreted as measuring the rate of inflation ([46], Sections 14.2–14.4). The inflation data are shown in the upper-left panel of Figure 4; there are n = 244 observations.

Figure 4

Top row: log-differenced CPI data and estimated kpacf of s-vine copula process using Gumbel copula sequence. Middle row: QQ-plots for residuals from models based on Gaussian (left) and Gumbel (right) copula sequences. Bottom row: QQ-plots of the data against fitted normal (left) and skewed Student (right) marginal distributions.

To establish a baseline model, we use an automatic ARMA selection algorithm, and this selects an ARMA(5,1) model. We first address the issue of whether the implied Gaussian copula sequence in an ARMA(5,1) model can be replaced by Gumbel, Clayton, Frank, or Joe copula sequences (or 180 degree rotations thereof); for any lag k at which the estimated kpacf τ k is negative, we retain a Gaussian copula and so the non-Gaussian copula sequences are actually hybrid sequences with some Gaussian terms. The data ( x 1 , , x n ) are transformed to pseudo-observations ( u 1 , , u n ) on the copula scale using the empirical distribution function, and the s-vine copula process is estimated by maximum-likelihood; this is the commonly used pseudo-maximum-likelihood method [12,19].

The best model results from replacing Gaussian copulas with Gumbel copulas, and the improvements in AIC and BIC are shown in the upper panel of Table 1; the improvement in fit is strikingly large. While the presented results relate to infinite-order processes, we note that very similar result (not tabulated) are obtained by fitting s-vine copula processes of finite order, where the kpacf is truncated at lag 30. Parameter estimates for the infinite-order models are presented in Table 2.

Table 1

Comparison of models by AIC and BIC: the top two lines relate to models for the pseudo-copula data ( u 1 , , u n ) , while the bottom three lines relate to full models of the original data ( x 1 , , x n )

No. pars AIC BIC
Gaussian copula process 6 −184.62 −163.64
Gumbel copula process 6 −209.28 −188.30
Gaussian process 8 372.73 400.71
Gaussian copula process + skewed Student margin 10 352.50 387.47
Gumbel copula process + skewed Student margin 10 319.17 354.14
Table 2

Parameter estimates and standard errors for s-vine copula processes with Gaussian and Gumbel copula sequences fitted to the pseudo-copula data ( u 1 , , u n )

θ (Ga) s.e. θ (Gu) s.e.
ϕ 1 −0.381 0.104 −0.232 0.130
ϕ 2 0.144 0.081 0.136 0.094
ϕ 3 0.197 0.063 0.180 0.061
ϕ 4 0.462 0.075 0.410 0.077
ϕ 5 0.324 0.063 0.266 0.061
ψ 1 0.870 0.098 0.771 0.118

The residual QQ-plots in the middle row of Figure 4 give further insight into the improved fit of the process with Gumbel copulas. In the usual manner, residuals are reconstructions of the unobserved innovation variables. If ( R ^ k ) k N denotes the sequence of estimated Rosenblatt forward functions, implied by the sequence ( C ^ k ) k N of estimated copulas, then residuals ( z 1 , , z n ) are constructed by setting z 1 = u 1 and z t = R ^ t 1 ( u [ 1 , t 1 ] , u t ) for t > 1 . To facilitate graphical analysis, these are transformed onto the standard normal scale so that the QQ-plots in the middle row of Figure 4 relate to the values ( Φ 1 ( z 1 ) , , Φ 1 ( z n ) ) and are against a standard normal reference distribution. The residuals from the baseline Gaussian copula appear to deviate from normality, whereas the residuals from the Gumbel copula model are much better behaved; the latter pass a Shapiro-Wilk test of normality ( p -value = 0.97), whereas the former do not ( p -value = 0.01).

The picture of the kpacf in the top right panel of Figure 4 requires further comment. This plot attempts to show how well the kpacf of the fitted copula sequence matches the empirical Kendall partial autocorrelations of the data. The continuous line is the kpacf of the Gumbel/Gaussian copula sequence used in the best-fitting vine copula model of ( u 1 , , u n ) . The vertical bars show the empirical Kendall partial autocorrelations of the data at each lag k . However, the method should really be considered as “semi-empirical” as it uses the fitted parametric copulas at lags 1 , , k 1 in order to construct the necessary data for lag k . The data used to estimate an empirical lag k rank correlation are the points

{ ( R ^ k 1 ( 2 ) ( u [ j k + 1 , j 1 ] , u j k ) , R ^ k 1 ( u [ j k + 1 , j 1 ] , u j ) ) , j = k + 1 , , n } ,

where R k ^ and R ^ k ( 2 ) denote the estimates of forward and backward Rosenblatt functions; it may be noted that these data are precisely the points at which the copula density c k is evaluated when the model likelihood based on c ( n ) in Eq. (5) is maximized.

The kpacf shows positive dependence between inflation rates at the first 5 lags; moreover, the choice of Gumbel copula suggests asymmetry and upper tail dependence in the bivariate distribution of inflation rates at time points that are close together; in other words, large values of inflation are particularly strongly associated with large values of inflation in previous quarters, while low values are more weakly associated.

We next consider composite models for the original data ( x 1 , , x n ) consisting of a marginal distribution and an s-vine copula process. The baseline model is simply a Gaussian process with Gaussian copula sequence and Gaussian marginal distribution. We experimented with a number of alternatives to the normal marginal and obtained good results with the skewed Student distribution from the family of skewed distributions proposed by Fernandez and Steel [18]. Table 1 contains results for models which combine the Gaussian and Gumbel copula sequences with the skewed Student margin; the improvement obtained by using a Gumbel sequence with a skewed Student margin is clear from the AIC and BIC values. The QQ-plots of the data against the fitted marginal distributions in the bottom row of Figure 4 also show the superiority of the skewed Student to the Gaussian distribution for this dataset.

The fitting method used for the composite model results in Table 1 is the two-stage IFM (inference functions for margins) method [25] in which the margin is estimated first, the data are transformed to approximately uniform using the marginal model, and the copula process is estimated by ML in a second step.

The estimated values of the degree of freedom and skewness parameters in the skewed Student t marginal distribution are ν = 3.19 and γ = 1.47 , respectively. These suggest that inflation rates (changes in log CPI) follow a heavy tailed, infinite-kurtosis distribution (tail index = 3.19) that is skewed to the right.

## 6 Conclusion

The s-vine processes provide a class of tractable stationary models that can capture non-linear and non-Gaussian serial dependence behaviour as well as any continuous marginal behaviour. By defining models of infinite order and using the approach based on the Kendall partial autocorrelation function (kpacf), we obtain a very natural generalization of classical Gaussian processes, such as Gaussian ARMA or ARFIMA.

The models are straightforward to apply. The parsimonious parametrization based on the kpacf makes maximum likelihood inference feasible. Analogues of many of the standard tools for time series analysis in the time domain are available, including estimation methods for the kpacf and residual plots that shed light on the quality of the fit of the copula model. By separating the issues of serial dependence and marginal modelling, we can obtain bespoke descriptions of both aspects that avoid the compromises of the more “off-the-shelf” classical approach. The example of Section 5.3 indicates the kind of gains that can be obtained; it seems likely that many empirical applications of classical ARMA could be substantially improved by the use of models in the general s-vine class. In combination with v-transforms [33], s-vine models could also be used to model data showing stochastic volatility following the approach developed by Bladt and McNeil [9].

To increase the practical options for model building it would be of interest to consider how copulas with more than one parameter, such as the t copula or the symmetrized Joe-Clayton copula [37] could be incorporated into the methodology. The parameters would have to be allowed to change in a smooth parsimonious manner such that the partial copula sequence ( C k ) k N converged to the independence copula while the Kendall correlations ( τ k ) k N followed the chosen form of kpacf for every k . This is a topic for further research.

The approach we have adopted should also be of interest to theoreticians as there are a number of challenging open questions to be addressed. While we have proposed definitions of causality and invertibility for general s-vine processes, we currently lack a mathematical methodology for checking convergence of causal and invertible representations for sequences of non-Gaussian pair copulas.

There are some very interesting questions to address about the relationship between the partial copula sequence ( C k ) k N , the rate of convergence of causal representations and the rate of ergodic mixing of the resulting processes. The example of Figure 1 indicates that, even for a finite-order process, some very extreme models can be constructed that mix extremely slowly. Moreover, Example 5 suggests that non-Gaussian copula sequences serve to further elongate memory in long-memory processes, and this raises questions about the effect of the tail dependence properties of the copula sequence on rates of convergence and length of memory.

It would also be of interest to confirm our conjecture that the pragmatic approach adopted in Section 5.2, in which the kpacf of the (infinite) partial copula sequence ( C k ) k N is matched to that of a stationary and ergodic Gaussian process, always yields a stationary and ergodic s-vine model, regardless of the choice of copula sequence. However, for practical applications, the problem can be obviated by truncating the copula sequence at some large finite lag k , so that we are dealing with an ergodic Markov chain as shown in Section 3.

1. Conflict of interest: The authors declare no conflict of interest.

2. Data availability statement: The analyses were carried out using R and the tscopula CRAN package. Code to reproduce the analyses may be found at https://github.com/ajmcneil/papers.

## Appendix

### A.1 Proof of Proposition 1

In this proof, we use the notation ( u ) i to denote the i th component of a vector u and u i to denote the vector u with i th component removed. An exchangeable copula satisfies C k ( u , v ) = C k ( v , u ) for all u , v , and hence, h k ( 2 ) ( u , v ) = h k ( 1 ) ( v , u ) . From this, it follows that R 1 ( 2 ) ( u , x ) = h 1 ( 2 ) ( x , u ) = h 1 ( 1 ) ( u , x ) = R 1 ( u , x ) . Part (i) follows by induction using the facts that for u = ( u 1 , , u k ) ( 0 , 1 ) k , we have ( u ¯ ) k = u 1 , u ¯ k = ( u k , , u 2 ) and u ¯ k ¯ = u 1 . We have that

R k ( 2 ) ( u ¯ , x ) = h k ( 2 ) ( R k 1 ( 2 ) ( u ¯ k , x ) , R k 1 ( u ¯ k , ( u ¯ ) k ) ) = h k ( 1 ) ( R k 1 ( u ¯ k , u 1 ) , R k 1 ( 2 ) ( u ¯ k , x ) ) = h k ( 1 ) ( R k 1 ( 2 ) ( u 1 , u 1 ) , R k 1 ( u 1 , x ) ) = R k ( u , x ) .

For part (ii), we observe that for any k > 1 the implication of part (i) is that for any t , s N the conditional distribution of U t + k U t + 1 = u 1 , , U t + k 1 = u k 1 is the same as that of U s + 1 U s + 2 = u k 1 , , U s + k = u 1 . It easily follows that ( U t + 1 , , U t + k ) = d ( U s + k , , U s + 1 ) , which proves reversibility of the process.

### A.2 Proof of Proposition 2

If ( X t ) is a Gaussian process, its marginal distributions of all orders are multivariate Gaussian. The general d-vine copula decomposition in Eq. (1) can be applied to each n -dimensional marginal density. Since the conditional distributions of pairs ( X j k , X j ) given intermediate variables are bivariate Gaussian distributions with covariance matrices that do not depend on the conditioning variables, the simplifying assumption holds for each pair copula density in Eq. (1) ([27], pp. 106–108). The stationarity assumption ensures that the joint density of the n -dimensional copula takes the form in Eq. (5).

Conversely, an s-vine process with Gaussian marginal density and Gaussian pair copulas is a stationary process with n -dimensional marginal densities of the form given in Eq. (7). These are the densities of multivariate Gaussian distributions, and the resulting process is a Gaussian process.

### A.3 Proof of Proposition 3

Let ( Z k ) k N be a sequence of iid standard uniform variables and ( U k ) k N a sequence of uniform random variables generated by setting U 1 = Z 1 and U k = R k 1 1 ( ( U 1 , , U k 1 ) , Z k ) for k > 1 , where ( R k ) k N denotes the sequence of Rosenblatt functions associated with the sequence of Gaussian pair copulas ( C k ) k N . Moreover, let ( X k ) k N be a sequence of standard Gaussian variables defined by setting X k = Φ 1 ( U k ) for all k .

It follows that, for any k 1 , ( X 1 , , X k + 1 ) N k + 1 ( 0 , P k + 1 ) , where P k + 1 is the ( k + 1 ) -dimensional correlation matrix implied by the acf ( ρ i ) i N of ( X i ) i N as used in Eq. 14. The standard result for the conditional distribution of a multivariate normal implies that

X k + 1 X 1 = x 1 , , X k = x k N ( ρ ¯ k P k 1 ( x 1 , , x k ) , 1 ρ k P k 1 ρ k ) ,

where ρ k = ( ρ 1 , , ρ k ) as in (14) and ρ ¯ k is the reversed vector. The mean of the conditional distribution is the best linear predictor of X k + 1 and the variance of the conditional distribution is the mean squared prediction error; let us write the former as j = 1 k ϕ j ( k ) x k + 1 j , where ϕ j ( k ) = ( P k 1 ρ k ) j , and the latter as σ k 2 . We then have

R k ( u , x ) = P ( U k + 1 x U 1 = u 1 , , U k = u k ) = P ( X k + 1 Φ 1 ( x ) X 1 = Φ 1 ( u 1 ) , , X k = Φ 1 ( u k ) ) = Φ Φ 1 ( x ) j = 1 k ϕ j ( k ) Φ 1 ( u k + 1 j ) σ k .

The expression σ k 2 = j = 1 k ( 1 α j 2 ) and the recursive formula in Eq. (18) for the coefficients ϕ j ( k ) follow from the Durbin–Levinson Algorithm ([11], Proposition 5.2.1).

It follows from Eq. (17) that, for k > 0 ,

(A.1) U k + 1 = R k 1 ( ( U 1 , , U k ) , Z k + 1 ) = Φ σ k Φ 1 ( Z k + 1 ) + j = 1 k ϕ j ( k ) Φ 1 ( U k + 1 j ) ,

which may be written in terms of the variables ( X k ) k N as X k + 1 = σ k ε k + 1 + j = 1 k ϕ j ( k ) X k + 1 j where we introduce the further notation ε k = Φ 1 ( Z k ) for all k N . An inductive argument then shows that this may be written in the form X k + 1 = σ k ε k + 1 + j = 1 k ψ j ( k ) ε k + 1 j with the coefficients ψ j ( k ) as defined in Eq. (20). Eq. (19) then follows easily from Eq. (11).

### A.4 Proof of Theorem 1

As in the proof of Proposition 3, we introduce the notation ( X t ) t Z and ( ε t ) t Z , where X t = Φ 1 ( U t ) and ε t = Φ 1 ( Z t ) . For fixed k , the formulas U t = S k ( ( Z t k , , Z t 1 ) , Z t ) and Z t = R k ( ( U t k , , U t 1 ) , U t ) translate to

X t = j = 0 k ψ j ( k ) ε t j and ε t = j = 0 k