Skip to content
BY 4.0 license Open Access Published by De Gruyter Open Access December 21, 2022

Averaging principle for two-time-scale stochastic differential equations with correlated noise

  • Tao Jiang and Yancai Liu EMAIL logo
From the journal Open Mathematics

Abstract

This article is devoted to studying the averaging principle for two-time-scale stochastic differential equations with correlated noise. By the technique of multiscale expansion of the solution to the backward Kolmogorov equation and consequent elimination of variables, we obtain the Kolmogorov equation corresponding to the reduced simplified system. The approximation of the slow component of the original system to the solution of the corresponding averaged equation is in the weak sense. An example is also provided to illustrate our result.

1 Introduction

Almost all dynamical systems in sciences, such as materials science, fluid dynamics, climate dynamics, celestial mechanics, and radiophysics, have a time hierarchy where the components evolve at quite different rates. In other words, some components evolve rapidly, while others vary slowly. Moreover, all important and interesting systems are essentially nonlinear and disturbed by inner and outer noises. Thus, to explore the interactions between multiscale, nonlinearity and uncertainty, many authors propose a lot of different methods in which the theory of averaging principle stands out. The averaging principle provides an effective tool to explore the asymptotic behavior of the multiscale dynamical systems.

The primary target of the averaging principle is to find the averaged equation, when the scale parameter ε tends to 0, for the slow component x of original multiscale system, and then to demonstrate the approximation between x and the solution of this averaged equation.

The averaging principle was initiated by Bogoliubov and Mitropolsky [1] for ordinary differential equations and then this theory is developed to deal with partial differential equations (PDEs). It is well known that realistic models incorporate uncertainty as an indispensable ingredient. The first result in this direction for stochastic differential equations (SDEs) was obtained by Khasminskii [2], which demonstrated the correctness of averaging principle in a weak sense. Now, the averaging principle for stochastic systems is attracting more and more interest and has a wide range of applications, see, e.g., [3,4, 5,6,7, 8,9] and references therein for the SDEs and [10,11, 12,13,14, 15,16,17, 18,19,20, 21,22] and references therein for stochastic partial differential equations (SPDEs).

However, as far as we know, all these works assume that the noises driving fast and slow components are independent. No result has been obtained for stochastic dynamical systems with two timescales as the noises are correlated. In order to fill the gap in this field of research, we consider a stochastic slow-fast dynamical system where the noise driving the fast components is correlated with the noise driving the slow components. More exactly, we consider the following SDEs driven by correlated noises on Z :

d x d t = 1 ε f 1 ( x , y ) + f 2 ( x , y ) + α 1 ( x , y ) d B 1 d t + α 2 ( x , y ) d B 2 d t , x ( 0 ) = x 0 , ( 1.1 ) d y d t = 1 ε 2 g 1 ( x , y ) + 1 ε g 2 ( x , y ) + 1 ε β ( x , y ) d B 2 d t , y ( 0 ) = y 0 , ( 1.2 )

where the small parameter ε > 0 is the ratio between the fast component y and the slow component x . B 1 and B 2 are independent standard Brownian motions. We take Z = R d , thus x R l , y R d l , or x T l , y T d l , then we have Z = T d . T d denotes the d -dimensional unit torus.

This article is organized as follows. In Section 2, some notations, hypotheses, and useful results we need in later sections are introduced. We present the main result of this article in Section 3, and the proof of this article is given in Section 4. An example is given to illustrate our result in Section 5.

2 Preliminaries

For the convenience of the following description, we first introduce some notations. For two d × d matrices A = ( a i j ) , B = ( b i j ) denote the inner product by A : B = t r ( A T B ) = i , j a i j b i j . Note that S : T = S T : T = 1 2 ( S + S T ) : T , if matrix T is symmetric. For vectors a , b , c R d , define the outer product, which is a matrix, of a and b by ( a b ) c = ( b c ) a .

The drift and diffusion coefficients should satisfy some appropriate conditions to derive the existence and uniqueness of the solution to the SDEs (1.1) and (1.2), as is stated in the following result (Theorem 6.1 in [23]):

Lemma 2.1

Assume that f 1 , f 2 , α 1 , α 2 , g 1 , and g 2 are all globally Lipschitz on Z . z 0 = ( x 0 , y 0 ) is independent of the Brownian motions B 1 and B 2 , and

E z 0 2 < .

Then the SDEs (1.1) and (1.2) have a unique solution z = ( x , y ) C ( R + ; Z ) .

The backward Kolmogorov equation for equations (1.1) and (1.2) is

(2.1) v t = 1 ε 2 0 v + 1 ε 1 v + 2 v , for ( x , y , t ) X × Y × R + ,

(2.2) v = ϕ ( x ) , for ( x , y , t ) X × Y × { 0 } ,

where

(2.3) 0 g 1 y + 1 2 B ( x , y ) : y y ,

(2.4) 1 f 1 x + g 2 y + 1 2 α 2 β T : y x + 1 2 β α 2 T : x y ,

(2.5) 2 f 2 x + 1 2 A ( x , y ) : x x ,

with

(2.6) A ( x , y ) α 1 α 1 T + α 2 α 2 T ,

(2.7) B ( x , y ) β β T .

The significance of the backward Kolmogorov equation above is displayed in the proof of the main result of this work in Section 4. Roughly speaking, we first write the solution to the backward Kolmogorov equation above in multiple-scale expansion. Then we identify the PDE satisfied by the main term of the expansion. Through the correspondence between the SDE and its backward Kolmogorov equation, we can finally derive a simplified equation to approximate the dynamics of x .

The adjoint operator of 0 is denoted by 0 . We impose the following assumptions:

A1: B ( x , y ) is a strictly and uniformly positive-definite matrix;

A2: 0 satisfies the Fredholm alternative. To be exact, let 0 : H H be an operator in H . Then

(i) Either

(2.8)(2.9) 0 u = γ , 0 U = Γ ,

have unique solutions for every γ , Γ H ;

or

(ii) the corresponding homogeneous equations

(2.10)(2.11) 0 V 0 = 0 , 0 v 0 = 0 ,

satisfy

1 dim ( N ( 0 ) ) = dim ( N ( 0 ) ) < ,

where N ( 0 ) and N ( 0 ) denote the null spaces of 0 and 0 , respectively. In the latter case, then equations (2.8) and (2.9) have a solution iff

v 0 N ( 0 ) , ( γ , v 0 ) = 0

and

V 0 N ( 0 ) ( Γ , V 0 ) = 0 ;

A3 (Centering condition): Y f 1 ( x , y ) ρ ( y ; x ) d y = 0 , x X .

With the help of the assumption (A1), the following result (Theorem 6.16 in [23]) holds:

Lemma 2.2

Suppose 0 , 0 on T d are equipped with periodic boundary conditions and B ( x , y ) is a strictly positive-definite matrix, uniformly in z = ( x , y ) T d . Then the following hold

  • N ( 0 ) = span { 1 } ;

  • N ( 0 ) = span { ρ } , i n f z T d ρ ( z ) > 0 .

1 stands for all functions that are independent of y.

Remark 2.1

Here we explain the significance of the result above. 0 is viewed as a differential operator in y , with x being a parameter. For variable elimination in Section 4, the natural ergodicity assumption to make is that

(2.12) 0 1 ( y ) = 0 ,

(2.13) 0 ρ ( y ; x ) = 0 .

Thus, Lemma 2.2 validates this ergodicity assumption.

3 Main result

As preparation for the main result of this article, we define the cell problem as follows:

(3.1) 0 Φ ( x , y ) = f 1 ( x , y ) , Y Φ ( x , y ) ρ ( y ; x ) d y = 0 .

This is a PDE in y , with x as a parameter. The solution Φ ( x , y ) will be crucial to the derivation in the next section. By assumptions (A2) and (A3), the existence and uniqueness of the solution to the cell problem (3.1) is guaranteed.

Remark 3.1

The existence and uniqueness for solutions to equation (3.1) is more complicated, in the case the matrix B ( x , y ) is degenerate or Y = R d . Analogous results are still possible, however, in function space settings with appropriate decay properties. See [24, 25,26].

Then define ρ by

(3.2) ρ ( x ) = Y ρ ( y ; x ) ( f 2 ( x , y ) + x Φ ( x , y ) f 1 ( x , y ) + g 2 ( x , y ) y Φ ( x , y ) + α 2 ( x , y ) β T ( x , y ) : x y Φ ( x , y ) ) d y F 2 ( x ) + F 1 ( x )

and σ ( x ) by

(3.3) σ ( x ) σ ( x ) T = A 2 ( x ) + 1 2 ( A 1 ( x ) + A 1 ( x ) T ) ,

where

(3.4) A 1 ( x ) 2 Y ρ ( y ; x ) [ f 1 ( x , y ) Φ ( x , y ) + β α 2 T ( y Φ ) T ] d y ,

(3.5) A 2 ( x ) Y ρ ( y ; x ) A ( x , y ) d y .

Now we state our main result, the derivation of which is in the next section.

Theorem 3.1

For t in scale of O ( 1 ) , the solution x ( t ) of (1.1) can be approximated by the solution X ( t ) of the following SDE,

(3.6) d X d t = ρ ( X ) + σ ( X ) d W d t , X ( 0 ) = x 0 ,

as ε 0 , where ρ ( X ) , σ ( X ) are defined before, and W ( t ) stands for a standard Wiener process, i.e., a standard Brownian motion which is independent of U and V.

Note that ρ and σ in equation (3.6) for X both depend on f 0 in the x equation.

Remark 3.2

The big-oh notation O in the theorem above is defined as follows: f ( x ) = O ( g ( x ) ) as x x 0 if there exists a constant C such that f ( x ) C g ( x ) for all x sufficiently close to x 0 . Thus, the phrase “ t in scale of O ( 1 ) ” means that the aforementioned theorem holds in finite time interval.

Remark 3.3

Note that merely from the knowledge of σ ( X ) σ T ( X ) , we cannot determine the diffusion matrix σ ( X ) uniquely. Therefore, the limiting equation (3.6) can neither be uniquely determined by equation (3.3), a consequence of the well-known fact that different SDEs can have the same generator. As a result, the approximation of the solution of (1.1) by the solution of (3.6) is in the sense of weak convergence of probability measures.

4 Derivation

In this section, we give the derivation of Theorem 3.1. By multiscale expansion, we seek the solution of (2.1) with the form

(4.1) v ( x , y , t ) = v 0 ( x , y , t ) + ε v 1 ( x , y , t ) + ε 2 v 2 ( x , y , t ) + .

Substitute this expansion into (2.1) and equate the coefficients of equal powers in ε . The first three of the hierarchy of equations are

(4.2) O 1 ε 2 : 0 v 0 = 0 ,

(4.3) O 1 ε : 0 v 1 = 1 v 0 ,

(4.4) O ( 1 ) : 0 v 2 = v 0 t + 1 v 1 + 2 v 0 .

By (2.12), we can deduce, from equation (4.2), v 0 ( x , y , t ) is independent of y , i.e., v 0 = v 0 ( x , t ) . As for equation (4.3), the solvability condition is satisfied by assumption (A3). From the expression of (2.4), we have

1 v 0 = f 1 ( x , y ) x v 0 ( x , t ) .

Equation (4.3) becomes

(4.5) 0 v 1 = f 1 ( x , y ) x v 0 ( x , t ) .

The general solution of (4.5) has the form

(4.6) v 1 ( x , y , t ) = Φ ( x , y ) x v 0 ( x , t ) + Φ 1 ( x , t ) ,

since 0 can be viewed as a differential operator in y with x being a parameter.

We set the function Φ 1 to 0 because it plays no role in what follows. Thus, solution v 1 is represented as a linear operator acting on v 0 . This form for v 1 is a useful representation since we aim to find a closed equation for v 0 . Substituting for v 1 in (4.5) indicates that Φ is the solution to the cell problem (3.1). The assumptions (A2) and (A3) ensure the existence of the solution to the cell problem and the normalization condition, the second equation in (3.1), makes it unique. The right-hand side equation (4.4) becomes

v 0 t 2 v 0 1 ( Φ x v 0 ) .

Hence for each fixed x , solvability of (4.4) leads to

(4.7) v 0 t = Y 2 v 0 ( x , t ) ρ d y + Y ( y ; x ) 1 ( Φ x v 0 ( x , t ) ) ρ ( y ; x ) d y I 1 + I 2 .

Using the symbols we introduce, the first term on the right-hand side of the equation above is

(4.8) I 1 = Y ( f 2 x + 1 2 A ( x , y ) : x x ) v 0 ( x , t ) ρ ( y ; x ) d y = F 2 ( x ) x v 0 ( x , t ) + 1 2 A 2 ( x ) : x x v 0 ( x , t ) .

As for the term I 2 , note that

(4.9) 1 = f 1 x + g 2 y + 1 2 α 2 β T : y x + 1 2 β α 2 T : x y = f 1 x + g 2 y + β α 2 T : x y .

The last equality is from the definition of the inner product between matrices.

By direct calculation, we have

(4.10) f 1 x ( Φ x v 0 ) = f 1 Φ : x x v 0 + ( x Φ f 1 ) x v 0 ,

(4.11) g 2 y ( Φ x v 0 ) = ( y Φ g 2 ) x v 0 ,

and

(4.12) β α 2 T : x y ( Φ x v 0 ) = β α 2 T : [ ( x y Φ ) x v 0 ] + β α 2 T : [ ( y Φ ) T x x v 0 ] = ( β α 2 T : x y Φ ) x v 0 + [ β α 2 T ( y Φ ) T ] : x x v 0 .

The last equality holds when β α 1 T is symmetric. The first term of the right-hand side of the last equality is from the property A : ( B c ) = ( A : B ) c . The second term of the right-hand side of the last equality is from the property A : ( B C ) = ( A B ) : C while A , C are both symmetric matrices.

Hence, I 2 = I 3 + I 4 where

(4.13) I 3 Y [ x Φ f 1 + y Φ g 2 + ( β α 2 T : x y Φ ) ] x v 0 ( x , t ) ρ ( y ; x ) d y

and

(4.14) I 4 Y [ f 1 Φ + β α 2 T ( y Φ ) T ] : x x v 0 ( x , t ) ρ ( y ; x ) d y .

Thus,

(4.15) I 2 = F 1 ( x ) x v 0 ( x , t ) + 1 2 A 1 ( x ) : x x v 0 ( x , t ) .

By equations (4.7), (4.8), and (4.15), we obtain the following equation:

(4.16) v 0 t = ρ ( x ) x v 0 + 1 2 σ ( x ) σ ( x ) T : x x v 0 .

This is exactly the backward Kolmogorov equation corresponding to the reduced dynamics given in (3.5). This completes the proof.

Remark 4.1

We note that equation (1.2) for the fast component y contains only one noise term, but the method we exploit still applies in the case that the fast component is driven by two independent Gaussian white noises. To be more exact, consider the following SDEs:

d x d t = 1 ε f 1 ( x , y ) + f 2 ( x , y ) + α 1 ( x , y ) d B 1 d t + α 2 ( x , y ) d B 2 d t , ( 4.17 ) d y d t = 1 ε 2 g 1 ( x , y ) + 1 ε g 2 ( x , y ) + 1 ε β 1 ( x , y ) d B 1 d t + 1 ε β 2 ( x , y ) d B 2 d t . ( 4.18 )

The operators 1 in (2.4) should be replaced by

(4.19) 1 = f 1 x + g 2 y + 1 2 α 2 β 2 T : y x + 1 2 β 2 α 2 T : x y .

Meanwhile, B ( x , y ) in (2.3) is replaced by

(4.20) B ( x , y ) = β 1 β 1 T + β 2 β 2 T .

Consequently, (3.2) and (3.4) are replaced by

(4.21) ρ ( x ) = Y ( f 2 ( x , y ) + x Φ ( x , y ) f 1 ( x , y ) + g 2 ( x , y ) y Φ ( x , y ) + ( α 1 ( x , y ) β 1 T ( x , y ) + α 2 ( x , y ) β 2 T ( x , y ) ) : x y Φ ( x , y ) ) ρ ( y ; x ) d y F 2 ( x ) + F 1 ( x )

and

(4.22) A 1 ( x ) 2 Y [ f 1 ( x , y ) Φ ( x , y ) + ( β 1 α 1 T + β 2 α 2 T ) ( y Φ ) T ] ρ ( y ; x ) d y .

After all these changes, the simplified equation has the same form as (3.6).

5 Example

To illustrate our result, consider the following system in T 2 :

d x d t = 1 ε ( 1 y 2 ) x + x y + x d B 1 d t + y d B 2 d t , x ( 0 ) = x 0 , ( 5.1 ) d y d t = 1 ε 2 α y + 1 ε 2 α d B 2 d t , y ( 0 ) = y 0 . ( 5.2 )

Hence, we have the corresponding functions f 1 ( x , y ) = ( 1 y 2 ) x , f 2 ( x , y ) = x y , α 1 ( x , y ) = x , α 2 ( x , y ) = y , g 1 ( x , y ) = α y , g 2 ( x , y ) = 0 , β ( x , y ) = 2 α .

In this case, 0 = α y y + α 2 y 2 and the cell problem becomes

α y Φ y α 2 Φ y 2 = ( 1 y 2 ) x .

This equation has a unique centered solution

Φ ( y , x ) = 1 2 α ( 1 y 2 ) x .

It is well known that the second and fourth moments take values 1 and 3, respectively, under the standard normal distribution, by which we obtain

1 2 A 1 = Y ρ ( y ; x ) x 2 ( 1 y 2 ) 2 2 α 2 α x y 2 d y = 1 α x 2 2 α x , A 2 = Y ρ ( y ; x ) ( x 2 + y 2 ) d y = x 2 + 1 , F 1 = Y ρ ( y ; x ) x ( 1 y 2 ) 2 2 α 2 α y 2 d y = 1 α x 2 α , F 2 = Y x y ρ ( y ; x ) d y = 0 .

Thus, we derive the simplified equation as follows to approximate the slow component x ( t ) :

(5.3) d X d t = X α 2 α + ( 1 + 2 α ) X 2 8 α X + 1 d W d t ,

where W ( t ) is a standard Brownian motion.

6 Conclusion

In this article, we study the averaging principle for systems of SDEs with two widely separated time scales, which is driven by correlated noises. By means of perturbation expansion of the solution to the backward Kolmogorov equation and consequent elimination of variables, we obtain the backward equation corresponding to the reduced simplified system. The solution to reduced SDE (3.6) approximates the slow variable x ( t ) of the initial systems (1.1) and (1.2) in the weak sense.

  1. Funding information: Tao Jiang is supported by the National Natural Science Foundation of China (11901178). Yancai Liu is supported by the High-Level Personal Foundation of Henan University of Technology (2020BS049).

  2. Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.

  3. Conflict of interest: The authors state no conflict of interest.

  4. Data availability statement: Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.

References

[1] N. N. Bogoliubov and Y. A. Mitropolsky, Asymptotic Methods in the Theory of Non-linear Oscillations, Gordon and Breach Science Publishers, New York, 1961. Search in Google Scholar

[2] R. Z. Khasminskii, On the principle of averaging the Itô stochastic differential equations, Kibernetika 4 (1968), 260–279, (in Russian). Search in Google Scholar

[3] M. I. Freidlin and A. D. Wentzell, Long-time behavior of weakly coupled oscillators, J. Stat. Phys. 123 (2006), 1311–1337. 10.1007/s10955-006-9133-8Search in Google Scholar

[4] D. Givon, Strong convergence rate for two-time-scale jump-diffusion stochastic differential systems, Multiscale Model. Simul. 6 (2007), no. 2, 577–594. 10.1137/060673345Search in Google Scholar

[5] D. Givon, I. G. Kevrekidis, and R. Kupferman, Strong convergence of projective integration schemes for singular perturbed stochastic differential systems, Commun. Math. Sci. 4 (2006), 707–729. 10.4310/CMS.2006.v4.n4.a2Search in Google Scholar

[6] J. Golec, Stochastic averaging principle for systems with pathwise uniqueness, Stoch. Anal. Appl. 13 (1995), no. 3, 307–322. 10.1080/07362999508809400Search in Google Scholar

[7] J. Golec and G. Ladde, Averaging principle and systems of singularly perturbed stochastic differential equations, J. Math. Phys. 31 (1990), 1116–1123. 10.1063/1.528792Search in Google Scholar

[8] A. Y. Veretennikov, On the averaging principle for systems of stochastic differential equations, Mat. Sb. 69 (1991), no. 1, 271–284. 10.1070/SM1991v069n01ABEH001237Search in Google Scholar

[9] A. Y. Veretennikov, On large deviations in the averaging principle for SDEs with full dependence, Ann. Probab. 27 (1999), no. 1, 284–296. 10.1214/aop/1022677263Search in Google Scholar

[10] J. Bao, G. Yin, and C. Yuan, Two-time-scale stochastic partial differential equations driven by α-stable noises: averaging principles, Bernoulli 23 (2017), no. 1, 645–669. 10.3150/14-BEJ677Search in Google Scholar

[11] C.-E. Bréhier, Strong and weak orders in averaging for SPDEs, Stochastic Process. Appl. 122 (2012), no. 7, 2553–2593. 10.1016/j.spa.2012.04.007Search in Google Scholar

[12] S. Cerrai and M. I. Freidlin, Averaging principle for a class of stochastic reaction-diffusion equations, Probab. Theory Related Fields 144 (2009), 137–177. 10.1007/s00440-008-0144-zSearch in Google Scholar

[13] S. Cerrai and A Khasminkii type averaging principle for stochastic reaction-diffusion equations, Ann. Appl. Probab. 19 (2009), no. 3, 899–948. 10.1214/08-AAP560Search in Google Scholar

[14] S. Cerrai, Averaging principle for systems of reaction-diffusion equations with polynomial nonlinearities perturbed by multiplicative noise, SIAM J. Math. Anal. 43 (2011), no. 6, 2482–2518. 10.1137/100806710Search in Google Scholar

[15] H. Fu and J. Duan, An averaging principle for two-scale stochastic partial differential equations, Stoch. Dyn. 11 (2011), no. 2–3, 353–367. 10.1142/S0219493711003346Search in Google Scholar

[16] H. Fu and J. Liu, Strong convergence in stochastic averaging principle for two time-scales stochastic partial differential equations, J. Math. Anal. Appl. 384 (2011), no. 1, 70–86. 10.1016/j.jmaa.2011.02.076Search in Google Scholar

[17] H. Fu, L. Wan, and J. Liu, Strong convergence in averaging principle for stochastic hyperbolic-parabolic equations with twotime-scales, Stochastic Process Appl. 125 (2015), no. 8, 3255–3279. 10.1016/j.spa.2015.03.004Search in Google Scholar

[18] B. Pei, Y. Xu, and J. L. Wu, Two-time-scales hyperbolic-parabolic equations driven by Poisson random measures: existence, uniqueness and averaging principles, J. Math. Anal. Appl. 447 (2017), no. 1, 243–268. 10.1016/j.jmaa.2016.10.010Search in Google Scholar

[19] W. Wang and A. J. Roberts, Average and deviation for slow-fast stochastic partial differential equations, J. Differential Equations 253 (2012), no. 5, 1265–1286. 10.1016/j.jde.2012.05.011Search in Google Scholar

[20] J. Xu, Lp− strong convergence of the averaging principle for slow-fast SPDEs with jumps, J. Math. Anal. Appl. 445 (2017), no. 1, 342–373. 10.1016/j.jmaa.2016.07.058Search in Google Scholar

[21] J. Xu, Y. Miao, and J. Liu, Strong averaging principle for two-time-scale non-autonomous stochastic FitzHugh-Nagumo system with jumps, J. Math. Phys. 57 (2016), no. 9, 092704. 10.1063/1.4963173Search in Google Scholar

[22] J. Xu, Y. Miao, and J. Liu, Strong averaging principle for slow-fast SPDEs with Poisson random measures, Discrete Contin. Dyn. Syst. Ser. B 20 (2015), no. 7, 2233–2256. 10.3934/dcdsb.2015.20.2233Search in Google Scholar

[23] G. A. Pavliotis and A. M. Stuart, Multiscale Methods: Averaging and Homogenization, Springer, New York, 2008. Search in Google Scholar

[24] E. Pardoux and Yu. Verrtennikov, On the Poisson equation and diffusion approximation. I, Ann. Probab. 29 (2001), no. 3, 1061–1085. 10.1214/aop/1015345596Search in Google Scholar

[25] E. Pardoux and Yu. Verrtennikov, On the Poisson equation and diffusion approximation. II, Ann. Probab. 31 (2003), no. 3, 1166–1192. 10.1214/aop/1055425774Search in Google Scholar

[26] E. Pardoux and Yu. Verrtennikov, On the Poisson equation and diffusion approximation. III, Ann. Probab. 33 (2005), no. 3, 1111–1133. 10.1214/009117905000000062Search in Google Scholar

Received: 2022-09-18
Revised: 2022-11-27
Accepted: 2022-12-01
Published Online: 2022-12-21

© 2022 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 1.12.2023 from https://www.degruyter.com/document/doi/10.1515/math-2022-0538/html
Scroll to top button