Applied Numerical Mathematics 57 (2007) 1097–1107 www.elsevier.com/locate/apnum Continuous parallel-iterated RKN-type PC methods for nonstiff IVPs Nguyen Huu Cong a,∗ , Nguyen Van Minh b a Faculty of Mathematics, Mechanics and Informatics, Hanoi University of Science, Vietnam b Faculty of Natural Science, Thai Nguyen University, Vietnam Available online 15 November 2006 Abstract This paper investigates parallel predictor–corrector (PC) iteration schemes based on direct collocation Runge–Kutta–Nyström (RKN) corrector methods with continuous output formulas for solving nonstiff initial-value problems (IVPs) for systems of special second-order differential equations y (t) = f(t, y(t)) Consequently, the resulting parallel-iterated RKN-type PC methods are provided with continuous output formulas The continuous numerical approximations are also used for predicting the stage values in the PC iteration processes In this way, we obtain parallel PC methods with continuous output formulas and high-order predictors Applications of the resulting parallel PC methods to a few widely-used test problems reveal that these new parallel PC methods are much more efficient when compared with the parallel-iterated RKN (PIRKN) methods and the sequential ODEX2 and DOPRIN codes from the literature © 2006 IMACS Published by Elsevier B.V All rights reserved Keywords: Runge–Kutta–Nyström methods; Predictor–corrector methods; Stability; Parallelism Introduction The arrival of parallel computers influences the development of numerical methods for the numerical solution of nonstiff initial-value problems (IVPs) for the systems of special second-order, ordinary differential equations (ODEs) y (t) = f t, y(t) , y(t0 ) = y0 , y (t0 ) = y0 , t0 t T, (1.1) where y, f ∈ Rd Among various numerical methods proposed so far, the most efficient methods for solving these problems are the explicit Runge–Kutta–Nyström (RKN) methods In the literature, sequential explicit RKN methods up to order 10 can be found in e.g., [16–20,22,23] In order to exploit the facilities of parallel computers, several class of parallel predictor–corrector (PC) methods based on RKN corrector methods have been investigated in e.g., [3–11,14,15,28,12,13] A common challenge in the latter-mentioned papers is to reduce, for a given order of accuracy, the required number of sequential f-evaluations per step, using parallel processors In the present paper, we investigate a particular class of parallel-iterated RKN-type PC methods based on direct collocation RKN corrector methods with continuous output formulas The continuous numerical approximations also are used as starting stage values in the * Corresponding author Current address: School of Graduate Studies, Vietnam National University, 144 Xuan Thuy, Cau Giay, Hanoi, Vietnam E-mail address: congnh@vnu.edu.vn (N.H Cong) 0168-9274/$30.00 © 2006 IMACS Published by Elsevier B.V All rights reserved doi:10.1016/j.apnum.2006.10.002 1098 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 PC iteration process In this way we obtain parallel PC methods that will be termed continuous parallel-iterated RKNtype PC methods (CPIRKN methods) Thus, we have achieved PC methods with dense output formulas and high-order predictors As a consequence, the resulting new CPIRKN methods require few numbers of sequential f-evaluations per step in the PC iteration process In Section 2, we shall consider RKN corrector methods with continuous output formulas (continuous RKN methods) Section formulates and investigates the CPIRKN methods, where the order of accuracy, the rate of convergence and the stability property are considered Furthermore, in Section 4, we present numerical comparisons of CPIRKN methods with traditional parallel-iterated RKN methods (PIRKN methods) and sequential numerical codes Continuous RKN methods A numerical method is inefficient, if the number of output points becomes very large (cf [24, p 188]) Therefore, in the literature, efficient numerical methods are often provided with a continuous output formula For constructing CPIRKN methods with such a continuous output formula in Section 3, in this section, we consider a continuous extension of implicit RKN methods Our starting point is an s-stage direct collocation (discrete) RKN method (see e.g., [4,12,25]) s Yn,i = un + hci un + h2 aij f(tn + cj h, Yn,j ), i = 1, , s, (2.1a) j =1 s un+1 = un + hun + h2 bj f(tn + cj h, Yn,j ), (2.1b) j =1 s un+1 = un + h dj f(tn + cj h, Yn,j ) (2.1c) j =1 Let us consider a continuous output formula defined by s un+ξ = un + hξ un + h2 bj (ξ )f(tn + cj h, Yn,j ) (2.1d) j =1 Here, in (2.1), ξ 2, un+ξ ≈ y(tn+ξ ), with tn+ξ = tn + ξ h, un+1 ≈ y(tn+1 ), un ≈ y(tn ), un+1 ≈ y (tn+1 ), un ≈ y (tn ) and h is the stepsize Furthermore, Yn,i , i = 1, , s, are the stage vector components representing numerical approximations to the exact solution values y(tn + ci h), i = 1, , s at nth step The s × s matrix A = (aij ), s-dimensional vectors b = (bj ), b(ξ ) = (bj (ξ )) and c = (cj ) are the method parameters in matrix or vector form The method defined by (2.1) will be called the continuous RKN method The step point and stage order of the (discrete) RKN method defined by (2.1a)–(2.1c) will be referred to as the step point and stage order of the continuous RKN method By the collocation principle, the continuous RKN corrector method (2.1) is of step point order p and stage order r both at least equal s (see [25]) This continuous RKN method can be conveniently presented by the Butcher tableau (see e.g., [25,2]) c yn+1 yn+1 yn+ξ A bT dT T b (ξ ) The matrix A and the vectors b and d are defined by the order conditions (see e.g., [12,25]) They can be explicitly expressed in terms of the collocation vector c as (cf [12]) A = P R −1 , bT = gT R −1 , d = gˆ S −1 (2.2) N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 1099 where j +1 ci j −1 j −1 , R = (rij ) = j ci , S = (sij ) = ci , j +1 1 , gˆ = (gˆ i ) = , i, j = 1, , s g = (gi ) = i+1 i P = (pij ) = The vector b(ξ ) in the continuous output formula (2.1d) is a vector function of ξ It satisfies the continuity conditions b(0) = and b(1) = b and will be determined by order conditions For the fixed stepsize h, these order conditions can be derived by replacing un+ξ , un and Yn,j in (2.1d) with the exact solution values and by requiring that the residue is of order s + in h Using Taylor expansions for sufficiently smooth function y(t) in the neighbourhood of tn , we obtain the order conditions for determining b(ξ ) D (j ) = ξ j +1 − bT (ξ )j cj −1 = 0, j +1 j = 1, , s (2.3a) The conditions (2.3a) can be seen to be of the form bT (ξ )R − gT diag ξ , ξ , , ξ s+1 = 0T (2.3b) From (2.3b) the explicit expression of the vector b(ξ ) is bT (ξ ) = gT diag ξ , ξ , , ξ s+1 R −1 (2.3c) In view of (2.2) and (2.3c), it follows that the continuity conditions for the vector b(ξ ) are clearly verified We have to note that the formula in (2.1b) is a special case of the continuous formula (2.1d) with ξ = It is evident that if the conditions (2.3) are satisfied, then we have the local order relation y(tn+ξ ) − un+ξ = O hs+2 (2.4) For the global order of continuous approximation defined by (2.1d) (continuous order), we have the following theorem: Theorem 2.1 If the function f is Lipschitz continuous and if the continuous RKN corrector method (2.1) is of step point order p, then the continuous output formula defined by (2.1d) gives rise to a continuous approximation of order (continuous order) p ∗ = min{p, s + 2} Proof Let us consider the global error estimate (without the local assumption: un = y(tn ), un = y (tn )) s y(tn+ξ ) − un+ξ = y(tn+ξ ) − un − hξ un − h2 bj (ξ )f(tn + cj h, Yn,j ) j =1 s = y(tn+ξ ) − y(tn ) − hξ y (tn ) − h2 bj (ξ )f tn + cj h, y(tn + cj h) j =1 + y(tn ) − un + hξ y (tn ) − un s + h2 bj (ξ ) f tn + cj h, y(tn + cj h) − f(tn + cj h, Yn,j ) (2.5) j =1 Since the function f is Lipschitz continuous, the following global order estimates hold: s y(tn+ξ ) − y(tn ) − hξ y (tn ) − h2 bj (ξ )f tn + cj h, y(tn + cj h) = O hs+2 , j =1 y(tn ) − un + hξ y (tn ) − un = O hp , s bj (ξ ) f tn + cj h, y(tn + cj h) − f(tn + cj h, Yn,j ) = O hs+2 h2 j =1 (2.6) 1100 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 Table Values of NCDp |NCDp∗ for problem (2.7) obtained by various continuous RKN methods Methods p p∗ Nstp = 200 Nstp = 400 Nstp = 800 Nstp = 1600 Nstp = 3200 Cont.Radau Cont.Gauss Cont.Radau Cont.Gauss 5 1.5|1.4 2.9|2.4 4.0|3.4 5.7|3.2 2.4|2.3 4.1|3.6 5.5|4.9 7.5|4.7 3.3|3.2 5.3|4.7 7.0|6.4 9.3|6.3 4.2|4.2 6.5|5.9 8.5|7.9 11.0|7.8 5.1|5.1 7.7|7.1 10.0|9.5 12.8|9.3 The proof of Theorem 2.1 follows from (2.5) and (2.6) In view of Theorem 2.1, if the step point order p of the continuous RKN method (2.1) is not less than s + 2, then the continuous order p ∗ of the approximation defined by (2.1d) is equal to s + ✷ Example 2.1 In order to show the order p ∗ of the continuous approximation (continuous order) as stated in Theorem 2.1, we consider continuous RKN methods based on direct collocation Radau IIA and Gauss–Legendre methods (see [4,25]) These methods will be called continuous Radau IIA (Cont.Radau) and continuous Gauss–Legendre (Cont.Gauss) We restrict our consideration to the 2-stage and 3-stage methods and apply them to the nonlinear Fehlberg problem (cf e.g., [16,17,19,20]): ⎛ −4t − 2 ⎞ y1 (t)+y2 (t) d2 y(t) ⎝ ⎠ y(t), = 2 dt −4t 2 y1 (t)+y2 (t) y(0) = (0, 1) , y (0) = (−2 π/2, 0)T , T π/2 t 10, (2.7) with highly oscillating exact solution given by y(t) = (cos(t ), sin(t ))T The absolute global error of the (discrete) approximation of order p obtained at tn+1 = and of the continuous approximation of order p ∗ obtained at tn+1.5 = tn + 1.5h = tn+1 + 0.5h are defined by y(tn+1 ) − yn+1 ∞ and y(tn+1.5 ) − yn+1.5 ∞ , respectively Table list the average numbers of correct decimal digits, i.e., the values defined by NCDp = − log10 y(tn+1 ) − yn+1 ∞ and by NCDp∗ = − log10 y(tn+1.5 ) − yn+1.5 ∞ The values NCDp |NCDp∗ listed in Table nicely show the theoretical orders p and p ∗ of the continuous RKN methods CPIRKN methods In this section, we consider a parallel PC iteration scheme based on the continuous RKN (corrector) methods This iteration scheme is given by s (0) (m) Yn,i = yn−1 + h(1 + ci )yn−1 + h2 bj (1 + ci )f tn−1 + cj h, Yn−1,j , i = 1, , s, (3.1a) j =1 s (k) aij f tn + cj h, Yk−1 n,j , Yn,i = yn + hci yn + h2 i = 1, , s, k = 1, , m, (3.1b) j =1 s (m) yn+1 = yn + hyn + h2 bj f tn + cj h, Yn,j , (3.1c) j =1 s dj f tn + cj h, Y(m) n,j , yn+1 = yn + h (3.1d) j =1 s (m) yn+ξ = yn + hξ yn + h bj (ξ )f tn + cj h, Yn,j j =1 (3.1e) N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 1101 Regarding (3.1a) as a predictor method and (2.1) as a corrector method, we arrive at a PC method in PE(CE)m E (m) mode Since the evaluations of f(tn−1 + cj h, Yn−1,j ), j = 1, , s, are available from the preceding step, we have in fact, a PC method in P (CE)m E mode In the PC method (3.1), the predictions (3.1a) are obtained by using continuous output formula (3.1e) from the (0) previous step If in (3.1a) we set Yn,i = yn + hci yn , i = 1, , s, the PC method (3.1a)–(3.1d) becomes the original parallel-iterated RKN methods (PIRKN methods) considered in [4,28] Therefore, we call the method (3.1), a con(k−1) tinuous parallel-iterated RKN-type PC method (CPIRKN method) Notice that the s components f(tn + cj h, Yn,j ), j = 1, , s, can be evaluated in parallel, provided that s processors are available, so that the number of sequential f-evaluations per step of length h in each processor equals s ∗ = m + Theorem 3.1 If the function f is Lipschitz continuous and if the continuous RKN corrector method (2.1) has step point order p, then the CPIRKN method (3.1) has step point order q = min{p, 2m + s + 2} and gives rise to a continuous approximation of order (continuous order) q ∗ = min{p, s + 2} Proof The proof of this theorem is very simple Suppose that f is Lipschitz continuous, yn = un = y(tn ) and yn = (0) un = y (tn ) Since Yn,i − Yn,i = O(hs+2 ) (see (2.4)) and each iteration raises the order of the iteration error by 2, we obtain the following order relations 2m+s+2 , Yn,i − Y(m) n,i = O h i = 1, , s, s bj f(tn + cj h, Yn,j ) − f tn + cj h, Y(m) n,j un+1 − yn+1 = h2 = O h2m+s+4 , j =1 s (m) un+1 − yn+1 = h dj f(tn + cj h, Yn,j ) − f tn + cj h, Yn,j = O h2m+s+3 (3.2) j =1 Hence, for the local truncation error of the CPIRKN method (3.1), we may write y(tn+1 ) − yn+1 = y(tn+1 ) − un+1 + [un+1 − yn+1 ] = O hp+1 + O h2m+s+4 , y (tn+1 ) − yn+1 = y (tn+1 ) − un+1 + [un+1 − yn+1 ] = O hp+1 + O h2m+s+3 (3.3) The order relations (3.3) gives the step point order q as stated in Theorem 3.1 for the CPIRKN method Furthermore, for the continuous order q ∗ of the continuous approximations defined by (3.1e), we may also write y(tn+ξ ) − yn+ξ = y(tn+ξ ) − un+ξ + [un+ξ − yn+ξ ] = y(tn+ξ ) − un+ξ + [un − yn ] + hξ [un − yn ] s (m) + h2 bj (ξ ) f(tn + cj h, Yn,j ) − f tn + cj h, Yn,j (3.4) j =1 From (3.2), (3.3) and Theorem 2.1 we have the following global order relations y(tn+ξ ) − un+ξ = O hmin{p,s+2} , un − yn = un − y(tn ) + y(tn ) − yn = O hmin{p,2m+s+2} hξ [un − yn ] = hξ un − y (tn ) + hξ y (tn ) − yn = O hmin{p+1,2m+s+3} s (m) bj (ξ ) f(tn + cj h, Yn,j ) − f tn + cj h, Yn,j h2 = O h2m+s+3 (3.5) j =1 The relations (3.4) and (3.5) then complete the proof of Theorem 3.1 ✷ Remark From Theorem 3.1, we see that by setting m = [(p − s − 1)/2] ([·] denoting the integer part function), we have a CPIRKN method of maximum step point order q = p (order of the corrector method) with minimum number of sequential f-evaluations per step s ∗ = [(p − s + 1)/2] 1102 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 3.1 Rate of convergence The rate of convergence of CPIRKN methods is defined by using the model test equation y (t) = λy(t), where λ runs through the eigenvalues of the Jacobian matrix ∂f/∂y (cf e.g., [4,6,7]) For this equation, we obtain the iteration error equation (j −1) (j ) Yn − Yn = zA Yn − Yn , z := h2 λ, j = 1, , m (3.6) Hence, with respect to the model test equation, the convergence rate is determined by the spectral radius ρ(zA) of the iteration matrix zA Requiring that ρ(zA) < 1, leads us to the convergence condition |z| < ρ(A) ρ(∂f/∂y)ρ(A) or h < (3.7) We shall call ρ(A) the convergence factor and 1/ρ(A) the convergence boundary of the CPIRKN method One can exploit the freedom in the choice of the collocation vector c of continuous RKN correctors for minimizing the convergence factor ρ(A), or equivalently, for maximizing the convergence region denoted by Sconv and defined as Sconv := z: |z| < 1/ρ(A) (3.8) The convergence factors ρ(A) for the CPIRKN methods used in the numerical experiments can be found in Section 3.2 Stability intervals The linear stability of the CPIRKN methods (3.1) is investigated by again using the model test equation y (t) = λy(t), where λ is assumed to be negative real By defining the matrix T B = b(1 + c1 ), , b(1 + cs ) , (0) (0) (0) for the model test equation, we can present the starting vector Yn = (Yn,1 , , Y1,s )T defined by (3.1a) in the form (m) Y(0) n = eyn−1 + h(e + c)yn−1 + zBYn−1 , where z := h2 λ Applying (3.1a)–(3.1c) to the model test equation yields (m−1) Y(m) n = eyn + chyn + zAYn = I + zA + · · · + (zA)m−1 (eyn + chyn ) + (zA)m Y(0) n (m) = zm+1 Am BYn−1 + I + zA + · · · + (zA)m−1 eyn + I + zA + · · · + (zA)m−1 chyn + zm Am eyn−1 + zm Am (e + c)hyn−1 , (3.9a) yn+1 = yn + hyn + zbT Y(m) n (m) = zm+2 bT Am BYn−1 + + zbT I + zA + · · · + (zA)m−1 e yn + + zbT I + zA + · · · + (zA)m−1 c hyn + zm+1 bT Am eyn−1 + zm+1 bT Am (e + c)hyn−1 , (3.9b) hyn+1 = hyn + zdT Y(m) n (m) = zm+2 dT Am BYn−1 + zdT I + zA + · · · + (zA)m−1 eyn + + zdT I + zA + · · · + (zA)m−1 c hyn + zm+1 dT Am eyn−1 + zm+1 dT Am (e + c)hyn−1 From (3.9) we are led to the recursion ⎛ (m) ⎞ ⎛ (m) ⎞ Yn Yn−1 ⎜ yn+1 ⎟ ⎜ yn ⎟ ⎜ ⎜ ⎟ ⎟ ⎜ hyn+1 ⎟ = Mm (z) ⎜ hyn ⎟ , ⎝ ⎝ ⎠ ⎠ yn yn−1 hyn hyn−1 (3.9c) (3.10a) N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 where Mm (z) is an (s + 4) × (s + 4) matrix defined by ⎛ m+1 m A B Pm−1 (z)e Pm−1 (z)c z ⎜ zm+2 bT Am B + zbT Pm−1 (z)e + zbT Pm−1 (z)c ⎜ Mm (z) = ⎜ zm+2 dT Am B zdT Pm−1 (z)e + zdT Pm−1 (z)c ⎝ T 0 0T 1103 ⎞ zm Am e zm Am (e + c) zm+1 bT Am e zm+1 bT Am (e + c) ⎟ ⎟ zm+1 dT Am e zm+1 dT Am (e + c) ⎟ , ⎠ 0 0 (3.10b) where Pm−1 (z) = I + zA + · · · + (zA)m−1 The matrix Mm (z) defined by (3.10) which determines the stability of the CPIRKN methods, will be called the amplification matrix, its spectral radius ρ(Mm (z)) the stability function For a given number of iterations m, the stability interval denoted by (−βstab (m), 0) of the CPIRKN methods is defined as −βstab (m), := z: ρ Mm (z) < 1, z We also call βstab (m) the stability boundary for a given m The stability boundaries βstab (m) for the CPIRKN methods used in the numerical experiments can be found in Section 4 Numerical experiments This section will report numerical results for the CPIRKN methods We confine our considerations to the CPIRKN methods with direct collocation continuous RKN corrector methods based on symmetric collocation vector c investigated in [6,11] The continuous s-stage RKN corrector methods (2.1) based on these symmetric collocation vectors have the orders p = p ∗ equal s + or s depending on whether s is odd or even (cf [6] and Theorem 2.1 in this paper) The symmetric collocation vectors were chosen such that the spectral radius ρ(A) of the RKN metrix A is minimized, so that the CPIRKN methods defined by (3.1) have “optimal” rate of convergence (see [6,11]) Table below lists the stability boundaries of the CPIRKN methods with continuous RKN corrector methods based on symmetric collocation vectors considered in [6,11] with s = 3, 4, 5, and with corresponding orders p = 4, 4, 6, The associated CPIRKN methods based on s-stage, p-order continuous RKN corrector methods will be denoted by CPIRKNsp For s = 3, 4, 5, and p = 4, 4, 6, we have the methods CPIRKN34, CPIRKN44, CPIRKN56 and CPIRKN66, respectively We observe that the stability boundaries of these CPIRKN methods show a rather irregular behaviour, however they are sufficiently large for nonstiff IVPs of the form (1.1) In the following, we shall compare the above CPIRKN methods with explicit parallel RKN methods and sequential codes from the literature For the CPIRKN methods, in the first step, we always use the trivial predictions given by (0) Y0,i = y0 + hci y0 , i = 1, , s The absolute error obtained at the end point of the integration interval is presented in the form 10−NCD (NCD may be interpreted as the average number of correct decimal digits) The computational efforts are measured by the values of Nseq denoting the total number of sequential f-evaluations required over the total number of integration steps denoted by Nstp Ignoring load balancing factors and communication times between processors in parallel methods, the comparison of various methods in this section is based on Nseq and the obtained NCDs The numerical experiments with small widely-used test problems taken from the literature below show a potential superiority in a parallel setting of the new CPIRKN methods over extant methods This superiority will be significant in a parallel machine if the test problems Table Stability boundaries βstab (m) for various CPIRKN methods Methods CPIRKN34 CPIRKN44 CPIRKN56 CPIRKN66 βstab (1) βstab (2) βstab (3) βstab (4) βstab (5) βstab (6) 1.472 0.087 2.367 0.535 2.039 9.765 3.114 0.184 0.424 6.701 1.236 2.051 0.075 1.579 0.617 1.582 9.869 3.417 0.155 5.996 0.790 5.926 2.309 6.031 1104 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 are large enough and/or the f-evaluations are expensive (cf e.g., [1]) In order to see the convergence behaviour of our CPIRKN methods, we follow a dynamical strategy in all PC methods for determining the number of iterations in the successive steps It seems natural to require that the iteration error is of the same order in h as the local error of the corrector This leads us to the stopping criterion (cf e.g., [3,4,6–8,10]) (m−1) Y(m) n − Yn ∞ TOL = Chp , (4.1) where C is a problem- and method-dependent parameter, p is the step point order of the corrector method All the computations were carried out on a 15-digit precision computer An actual implementation on a parallel machine is a subject of further studies 4.1 Comparison with parallel methods We shall report numerical results obtained by PIRKN methods, ones of the best parallel explicit RKN methods available in the literature proposed in [4,28], and the CPIRKN methods considered in this paper We consider indirect PIRKN (Ind.PIRKN) methods investigated in [28] and direct PIRKN (Dir.PIRKN) methods investigated in [4] We select a test set of three problems taken from the literature These three problems possess exact solutions in closed form Initial conditions are taken from the exact solutions 4.1.1 Linear nonautonomous problem As a first numerical test, we apply the various p-order PC methods to the linear nonautonomous problem (cf e.g., [4,6,7]) d2 y(t) = dt −2α(t) + 2(α(t) − 1) −α(t) + y(t), α(t) − α(t) = max cos2 (t), sin2 (t) , y(0) = (0, 0)T , t 20, y (0) = (−1, 2)T , (4.2) y(t) = (− sin(t), sin(t))T with exact solution The numerical results listed in Table clearly show that the CPIRKN methods are much more efficient than the indirect and direct PIRKN methods of the same order For this linear problem, all the CPIRKN methods need only about one iteration per step Notice that because of round-off errors, we cannot expect 15 digits accuracy for the numerical results As a consequence, Table contains four empty spots in the last two lines where the numerical results were in the neighbourhood of the accuracy-limits of the computer and therefore considered as unreliable 4.1.2 Nonlinear Fehlberg problem For the second numerical test, we apply the various p-order PC methods to the well-known nonlinear Fehlberg problem (2.7) considered in Section The numerical results are reported in Table These numerical results show that the CPIRKN methods are again by far superior to the indirect and direct PIRKN methods of the same order For this nonlinear Fehlberg problem, the number of iterations m needed at each step for all CPIRKN methods is one or two Table Values of NCD/Nseq for problem (4.2) obtained by various p-order parallel PC methods PC methods p Nstp = 80 Nstp = 160 Nstp = 320 Nstp = 640 Nstp = 1280 C Ind.PIRKN Dir.PIRKN 4 4.0/239 5.2/239 5.3/480 6.4/479 6.5/960 7.6/960 7.7/1920 8.8/1920 8.9/3840 10.0/3840 10−1 10−1 CPIRKN34 CPIRKN44 4 5.6/161 5.8/161 6.9/321 7.0/321 8.1/641 8.2/641 9.3/1281 9.4/1281 10.5/2561 10.6/2561 10−1 10−1 Ind.PIRKN Dir.PIRKN 6 7.4/360 8.0/354 9.2/721 9.9/710 11.0/1441 11.7/1420 12.8/2881 13.5/2839 14.6/5769 15.3/5678 10−3 10−3 CPIRKN56 CPIRKN66 6 9.8/173 10.2/162 11.7/322 11.9/322 13.8/642 13.9/642 10−3 10−3 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 1105 Table Values of NCD/Nseq for problem (2.7) obtained by various p-order parallel PC methods PC methods p Nstp = 200 Nstp = 400 Nstp = 800 Nstp = 1600 Nstp = 3200 C Ind.PIRKN Dir.PIRKN 4 1.7/728 2.4/722 2.8/1457 3.6/1445 4.0/2915 4.8/2889 5.2/5829 6.0/5778 6.5/11658 7.2/11555 102 102 CPIRKN34 CPIRKN44 4 3.3/523 3.3/473 4.6/1007 4.5/866 5.8/1942 5.7/1601 7.0/3713 6.9/3201 8.2/7033 8.1/6401 102 102 Ind.PIRKN Dir.PIRKN 6 4.0/900 5.0/896 5.8/1812 6.8/1807 7.6/3625 8.6/3615 9.4/7247 10.4/7230 11.2/14496 12.2/14458 103 103 CPIRKN56 CPIRKN66 6 6.5/526 6.7/468 8.3/999 8.5/878 10.1/1941 10.3/1611 11.9/3763 12.1/3202 13.7/7254 13.9/6402 103 103 Table Values of NCD/Nseq for problem (4.3) obtained by various p-order parallel PC methods PC methods p Nstp = 100 Nstp = 200 Nstp = 400 Nstp = 800 Nstp = 1600 C Ind.PIRKN Dir.PIRKN 4 2.9/229 2.8/229 3.7/600 4.9/600 4.9/1200 6.2/1200 6.1/2400 7.4/2400 7.3/4800 8.6/4800 101 101 CPIRKN34 CPIRKN44 4 4.6/201 3.3/201 5.8/401 4.5/401 6.9/801 5.7/801 8.1/1601 6.9/1601 9.3/3201 8.1/3201 101 101 Ind.PIRKN Dir.PIRKN 6 5.0/400 5.8/400 6.8/400 7.5/800 8.6/1600 9.3/1600 10.4/3200 11.1/3200 12.2/6400 12.9/6400 10−1 10−1 CPIRKN56 CPIRKN66 6 7.8/227 6.4/210 9.2/440 8.2/402 10.8/831 10.0/802 12.7/1602 11.7/1602 14.5/3202 13.6/3202 10−1 10−1 4.1.3 Newton’s equation of motion problem The third numerical example is the two-body gravitational problem for Newton’s equation of motion (see [27, p 245]) d2 y1 (t) y1 (t) =− , dt ( y12 (t) + y22 (t))3 y2 (t) d2 y2 (t) =− , d t ( y12 (t) + y22 (t))3 y1 (0) = − ε, y1 (0) = 0, y2 (0) = 0, y2 (0) = 1+ε 1−ε t 20, (4.3) This problem can also√be found in [20] or in the test set of problems in [26] The solution components are y1 (t) = cos(u(t))−ε, y2 (t) = (1 + ε)(1 − ε) sin(u(t)), where u(t) is the solution of Kepler’s equation t = u(t)−ε sin(u(t)) and ε denotes the eccentricity of the orbit In this example, we set ε = 0.3 The numerical results for this problem are given in Table and give rise to nearly the same conclusions as formulated in the two previous examples 4.2 Comparison with sequential codes In Section 4.1, the CPIRKN methods were compared with indirect and direct PIRKN methods In this section, we shall compare these CPIRKN methods with some of the best sequential codes currently available We restricted the numerical experiments to the comparison of two of our 6-order CPIRKN56 and CPIRKN66 methods with two well-known sequential codes for the nonlinear Fehlberg problem (2.7), that is the codes DOPRIN and ODEX2 taken from [24] We reproduced the best results obtained by these sequential codes given in the literature (cf e.g., [28,9]) and added the results obtained by CPIRKN56 and CPIRKN66 methods In spite of the fact that the results of the sequential codes are obtained using a stepsize strategy, whereas CPIRKN56 and CPIRKN66 methods are applied with fixed stepsizes, it is the CPIRKN56 and CPIRKN66 methods that are the most efficient (see Table 6) 1106 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 Table Comparison with sequential codes for problem (2.7) Codes/Methods Nstp NCD 49 53 43 57 81 2.7 4.8 6.5 8.8 10.9 746 1122 1493 2039 2907 DOPRIN (from [24]) 79 353 1208 4466 3.8 8.3 12.3 16.3 633 2825 9665 35729 CPIRKN56 (in this paper) 200 400 800 1600 3200 6.5 8.3 10.1 11.9 13.7 526 999 1941 3763 7254 CPIRKN66 (in this paper) 200 400 800 1600 3200 6.7 8.5 10.3 12.1 13.9 468 878 1611 3202 6402 ODEX2 (from [24]) Nseq Concluding remarks In this paper, we proposed a new class of parallel PC methods called continuous parallel-iterated RKN-type PC methods (CPIRKN methods) based on continuous RKN corrector methods Three numerical experiments showed that the CPIRKN methods are much superior to the well-known PIRKN methods and ODEX2 and DOPRIN codes available in the literature The paper limits its focus to IVPs of the form y (t) = f(t, y(t)), y(t0 ) = y0 , y (t0 ) = y0 , however there has been proposed RKN methods for the more general problem y (x) = f(t, y(t), y (t)), y(t0 ) = y0 , y (t0 ) = y0 (see e.g., [21]) In a forthcoming paper, we will extend the ideas of this paper to this more general problem References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] K Burrage, Parallel and Sequential Methods for Ordinary Differential Equations, Clarendon Press, Oxford, 1995 J.C Butcher, The Numerical Analysis of Ordinary Differential Equations, Runge–Kutta and General Linear Methods, Wiley, New York, 1987 N.H Cong, An improvement for parallel-iterated Runge–Kutta–Nyström methods, Acta Math Viet 18 (1993) 295–308 N.H Cong, Note on the performance of direct and indirect Runge–Kutta–Nyström methods, J Comput Appl Math 45 (1993) 347–355 N.H Cong, Direct collocation-based two-step Runge–Kutta–Nyström methods, SEA Bull Math 19 (1995) 49–58 N.H Cong, Explicit symmetric Runge–Kutta–Nyström methods for parallel computers, Comput Math Appl 31 (1996) 111–122 N.H Cong, Explicit parallel two-step Runge–Kutta–Nyström methods, Comput Math Appl 32 (1996) 119–130 N.H Cong, RKN-type parallel block PC methods with Lagrange-type predictors, Comput Math Appl 35 (1998) 45–57 N.H Cong, Explicit pseudo two-step RKN methods with stepsize control, Appl Numer Math 38 (2001) 135–144 N.H Cong, N.T Hong Minh, Parallel block PC methods with RKN-type correctors and Adams-type predictors, Internat J Comput Math 74 (2000) 509–527 N.H Cong, N.T Hong Minh, Fast convergence PIRKN-type PC methods with Adams-type predictors, Internat J Comput Math 77 (2001) 373–387 N.H Cong, N.T Hong Minh, Parallel-iteratet pseudo two-step Runge–Kutta–Nyström methods for nonstiff second-order IVPs, Comput Math Appl 44 (2002) 143–155 N.H Cong, N.V Minh, Improved parallel-iteratet pseudo two-step RKN methods for nonstiff problems, SEA Bull Math., in press N.H Cong, K Strehmel, R Weiner, Runge–Kutta–Nyström-type parallel block predictor–corrector methods, Adv Comput Math 10 (1999) 115–133 N.H Cong, K Strehmel, R Weiner, A general class of explicit pseudo two-step RKN methods on parallel computers, Comput Math Appl 38 (1999) 17–30 N.H Cong, N Van Minh / Applied Numerical Mathematics 57 (2007) 1097–1107 1107 [16] E Fehlberg, Klassische Runge–Kutta–Nyström Formeln mit Schrittweiten-Kontrolle für Differentialgleichungen x = f (t, x), Computing 10 (1972) 305–315 [17] E Fehlberg, Eine Runge–Kutta–Nyström Formel 9-ter Ordnung mit Schrittweitenkontrolle für Differentialgleichungen x = f (t, x), Z Angew Math Mech 61 (1981) 477–485 [18] E Fehlberg, S Filippi, J Gräf, Eine Runge–Kutta–Nyström Formelpaar der Ordnung 10(11) für Differentialgleichungen y = f (t, y), Z Angew Math Mech 66 (1986) 265–270 [19] S Filippi, J Gräf, Ein Runge–Kutta–Nyström Formelpaar der Ordnung 11(12) für Differentialgleichungen der Form y = f (t, y), Computing 34 (1985) 271–282 [20] S Filippi, J Gräf, New Runge–Kutta–Nyström formula-pairs of order 8(7), 9(8), 10(9) and 11(10) for differential equations of the form y = f (t, y), J Comput Appl Math 14 (1986) 361–370 [21] J.M Fine, Low order Runge–Kutta–Nyström methods, Computing 38 (1987) 281–297 [22] E Hairer, Méthodes de Nyström pour l’équation differentielle y (t) = f (t, y), Numer Math 27 (1977) 283–300 [23] E Hairer, A one-step method of order 10 for y (t) = f (t, y), IMA J Numer Anal (1982) 83–94 [24] E Hairer, S.P Nørsett, G Wanner, Solving Ordinary Differential Equations, I Nonstiff Problems, second revised ed., Springer, Berlin, 1993 [25] P.J van der Houwen, B.P Sommeijer, N.H Cong, Stability of collocation-based Runge–Kutta–Nyström methods, BIT 31 (1991) 469–481 [26] T.E Hull, W.H Enright, B.M Fellen, A.E Sedgwick, Comparing numerical methods for ordinary differential equations, SIAM J Numer Anal (1972) 603–637 [27] L.F Shampine, M.K Gordon, Computer Solution of Ordinary Differential Equations, the Initial Value Problems, W.H Freeman and Company, San Francisco, 1975 [28] B.P Sommeijer, Explicit, high-order Runge–Kutta–Nyström methods for parallel computers, Appl Numer Math 13 (1993) 221–240 ... Numerical Mathematics 57 (2007) 1097–1107 PC iteration process In this way we obtain parallel PC methods that will be termed continuous parallel-iterated RKNtype PC methods (CPIRKN methods) Thus,... parallel PC methods called continuous parallel-iterated RKN-type PC methods (CPIRKN methods) based on continuous RKN corrector methods Three numerical experiments showed that the CPIRKN methods. .. consider RKN corrector methods with continuous output formulas (continuous RKN methods) Section formulates and investigates the CPIRKN methods, where the order of accuracy, the rate of convergence