You are on page 1of 7

2012 IEEE 53rd Annual Symposium on Foundations of Computer Science

Quasi-optimal multiplication of linear differential operators

Alexandre Benoit Alin Bostan† Joris van der Hoeven‡


UPMC INRIA CNRS & École polytechnique
France France France
Email: Alexandre.Benoit@lip6.fr Email: Alin.Bostan@inria.fr Email: vdhoeven@lix.polytechnique.fr

Abstract—We show that linear differential oper- constitutes a first useful step towards understanding the
ators with polynomial coefficients over a field of bit complexity. Of course, in the special, very important
characteristic zero can be multiplied in quasi-optimal case when the field K is finite, both complexities coincide
time. This answers an open question raised by van
der Hoeven. up to a constant factor.
The complexities of operations in the rings K[x] and
Keywords-Linear differential operators; multiplica-
Kr×r have been intensively studied during the last
tion; algebraic algorithms; computational complexity.
decades. It is well established that polynomial multi-
plication is a commutative complexity yardstick, while
I. Introduction matrix multiplication is a non-commutative complexity
The product of polynomials and the product of matri- yardstick, in the sense that the complexity of oper-
ces are two of the most basic operations in mathematics; ations in K[x] (resp. in Kr×r ) can generally be ex-
the study of their computational complexity is central pressed in terms of the cost of multiplication in K[x]
in computer science. In this paper, we will be inter- (resp. in Kr×r ), and for most of them, in a quasi-linear
ested in the computational complexity of multiplying way [2], [4], [8], [24], [14].
two linear differential operators. These algebraic objects Therefore, understanding the algebraic complexity of
encode linear differential equations, and form a non- multiplication in K[x] and Kr×r is a fundamental ques-
commutative ring that shares many properties with tion. It is well known that polynomials of degrees < d
the commutative ring of usual polynomials [21], [22]. can be multiplied in time M(d) = O(d log d log log d)
The structural analogy between polynomials and linear using algorithms based on the Fast Fourier Transform
differential equations was discovered long ago by Libri (FFT) [11], [25], [9], and two r × r matrices in Kr×r
and Brassinne [18], [7], [13]. Yet, the algorithmic study can be multiplied in time O(rω ), with 2  ω  3
of linear differential operators is currently much less [27], [23], [12]. The current tightest upper bound, due to
advanced than in the polynomial case: the complexity Vassilevska Williams [28], is ω < 2.3727, following work
of multiplication has been addressed only recently [16], of Coppersmith and Winograd [12] and Stothers [26].
[6], but not completely solved. The aim of the present Finding the best upper bound on ω is one of the most
work is to make a step towards filling this gap, and to important open problems in algebraic complexity theory.
solve an open question raised in [16]. In a similar vein, our thesis is that understanding the
Let K be an effective field. That is, we assume data algebraic complexity of multiplication of linear differen-
structures for representing the elements of K and al- tial operators is a very important question, since the
gorithms for performing the field operations. The aim complexity of more involved, higher-level operations on
of algebraic complexity theory is to study the cost of linear differential operators can be reduced to that of
basic or more complex algebraic operations over K (such multiplication [17].
as the cost of computing the greatest common divisor From now on, we will assume that the base field K has
of two polynomials of degrees less than d in K[x], or characteristic zero. Let K[x, ∂] denote the associative al-
the cost of Gaussian elimination on an r × r matrix gebra Kx, ∂; ∂x = x∂ +1 of linear differential operators
d
in Kr×r ) in terms of the number of operations in K. in ∂ = dx with polynomial coefficients in x. Any element
The algebraic complexity usually does not coincide with L of K[x, ∂] can be written as a finite sum i Li (x)∂ i
the bit complexity, which also takes into account the for uniquely determined polynomials Li in K[x]. We say
potential growth of the actual coefficients in K. Never- that L has bidegree less than (d, r) in (x, ∂) if L has
theless, understanding the algebraic complexity usually degree less than r in ∂, and if all Li ’s have degrees less
than d in x. The degree in ∂ of L is usually called the
† Supported by the Microsoft Research – INRIA Joint Centre.
‡ Supported by the ANR-09-JCJC-0098-01 MaGiX project, as order of L.
well as by a Digiteo 2009-36HD grant and Région Île-de-France. The main difference with the commutative ring K[x, y]

0272-5428/12 $26.00 © 2012 IEEE 524


DOI 10.1109/FOCS.2012.57
of usual bivariate polynomials is the commutation rule In the important case d  r, our complexity bound
∂x = x∂ + 1 that simply encodes, in operator notation, reads Õ(rω−1 d). This is quasi-linear (thus quasi-optimal)
d d
Leibniz’s differentiation rule dx (xf ) = x dx (f ) + f . with respect to d. Moreover, by the equivalence result
This slight difference between K[x, ∂] and K[x, y] has from [6, §3], the exponent of r is also the best pos-
a considerable impact on the complexity level. On the sible. Besides, under the (plausible, still conjectural)
one hand, it is classical that multiplication in K[x, y] assumption that ω = 2, the complexity in Theorem 1
can be reduced to that of polynomials in K[x], due to a is almost linear with respect to the output size. For
technique commonly called Kronecker’s trick [19], [14]. r = 1 we retrieve the fact that multiplication in K[x] in
As a consequence, any two polynomials of degrees less degree < d can be done in quasi-linear time Õ(d); from
than d in x, and less than r in y, can be multiplied in this perspective, the result of Theorem 1 can be seen
quasi-optimal time O(M(dr)). On the other hand, under as a generalization of the fast multiplication for usual
our hypothesis that K has characteristic zero, it was polynomials.
shown by van der Hoeven [16] that the product of two In an expanded version [3] of this paper, we will
elements from K[x, ∂] of bidegree less than (n, n) can be show that analogues of Theorem 1 also hold for other
computed in time O(nω ). Moreover, it has been proved types of skew polynomials. More precisely, we will prove
in [6] that conversely, multiplication in Kn×n can be re- similar complexity bounds when the skew indeterminate
duced to a constant number of multiplications in K[x, ∂], ∂ : f (x) → f  (x) is replaced by the Euler derivative
in bidegree less than (n, n). In other words, multiplying δ : f (x) → xf  (x), or a shift operator σ c : f (x) →
operators of well-balanced bidegree is computationally f (x+c), or a dilatation χq : f (x) → f (qx). Most of these
equivalent to matrix multiplication. other cases are treated by showing that rewritings such
However, contrary to the commutative case, higher- as δ ↔ x∂ or σ c ↔ exp(c∂) can be performed efficiently.
level operations in K[x, ∂], such as the least common left We will also prove complexity bounds for a few other
multiple (LCLM) and the greatest common right divisor interesting operations on skew polynomials.
(GCRD), do not preserve well-balanced bidegrees [15], Main ideas. The fastest known algorithms for multipli-
[5]. For instance, the LCLM of two operators of bidegrees cation of usual polynomials in K[x] rely on an evaluation-
less than (n, n) is of bidegree less than (2n(n + 1), 2n) = interpolation strategy at special points in the base
O(n2 , n), and this bound is generically reached. This is field K [11], [25], [9]. This reduces polynomial multiplica-
a typical phenomenon: operators obtained from compu- tion to the “inner product” in K. We adapt this strategy
tations in K[x, ∂] tend to have much larger degrees in x to the case of linear differential operators in K[x, ∂]: the
than in ∂. evaluation “points” are exponential polynomials of the
In the general case of operators with possibly unbal- form xn eαx on which differential operators act nicely.
anced degrees d in x and r in ∂, the naive algorithm has With this choice, the evaluation and interpolation of
cost O(d2 r2 min(d, r)); a better algorithm, commonly operators is encoded by Hermite evaluation and interpo-
attributed to Takayama, has complexity Õ(dr min(d, r)). lation for usual polynomials (generalizing the classical
We refer to [6, §2] for a review of these algorithms. Lagrange interpolation), for which quasi-optimal algo-
When r  d  r4−ω , the best current upper bound for rithms exist. For operators of bidegree less than (d, r) in
multiplication is O(rω−2 d2 ) [16], [17]. It was asked by (x, ∂), with r  d, we use p = O(r/d) evaluation points,
van der Hoeven [16, §6] whether this complexity could and encode the inner multiplication step by p matrix
be lowered to Õ(rω−1 d). Here, and hereafter, the soft-O multiplications in size d. All in all, this gives an FFT-
notation Õ( ) indicates that polylogarithmic factors in d type multiplication algorithm for differential operators of
and in r are neglected. The purpose of the present work complexity Õ(dω−1 r). Finally, we reduce the case r < d
is to provide a positive answer to this open question. Our to the case r  d. To do this efficiently, we design a fast
main result is encapsulated in the following theorem: algorithm for the computation of the so-called reflection
Theorem 1: Let K be an effective field of characteristic of a differential operator, a useful ring morphism that
zero. Operators in K[x, ∂] of bidegree less than (d, r) in swaps the indeterminates x and ∂, and whose effect is
(x, ∂) can be multiplied using exchanging degrees and orders.
II. Preliminaries
Õ(min(d, r)ω−2 dr)
Recall that K denotes an effective field of character-
operations in K. istic zero. Throughout the paper, K[x]d will denote the
set of polynomials of degree less than d with coefficients
Actually, we will prove slightly more refined versions in the field K, and K[x, ∂]d,r will denote the set of linear
of this theorem (see Theorems 3 and 5 below), by making differential operators in K[x, ∂] with degree less than r
the hidden log-terms in the complexity explicit. in ∂, and polynomial coefficients in K[x]d .

525
The cost of our algorithms will be measured by the of roots of unity. It is also natural to represent the
number of field operations in K they use. We recall evaluation of L at a suitable number of such powers by
that polynomials in K[x]d can be multiplied within a matrix. More precisely, given k ∈ N, we may regard
M(d) = O(d log d log log d) = Õ(d) operations in K, L as an operator from K[x]k to K[x]k+d . We may also
using the FFT-based algorithms in [25], [9], and that regard elements of K[x]k and K[x]k+d as column vectors,
ω denotes a feasible exponent for matrix multiplication written in the canonical bases with powers of x. We will
over K, that is, a real constant 2  ω  3 such that two denote by
r × r matrices with coefficients in K can be multiplied ⎛ ⎞
L(1)0 ··· L(xk−1 )0
in time O(rω ). Throughout this paper, we will make ⎜ ⎟
the classical assumption that M(d)/d is an increasing Φk+d,k = ⎜ .. .. ⎟
L ⎝ . . ⎠
function in d.
Most basic polynomial operations in K[x]d (division, L(1)k+d−1 · · · L(x k−1
)k+d−1
Taylor shift, extended gcd, multipoint evaluation, inter- the matrix of the K-linear map L : K[x]k → K[x]k+d
polation, etc.) have cost Õ(d) [2], [4], [8], [24], [14]. Our with respect to these bases. Given two operators K, L in
algorithms will make a crucial use of the following result K[x, ∂]d,r , we clearly have
due to Chin [10], see also [20] for a formulation in terms
of structured matrices. Φk+2d,k
KL = Φk+2d,k+d
K Φk+d,k
L , for all k  0.
Theorem 2 (Fast Hermite evaluation–interpolation):
Let K be an effective field of characteristic zero, For k = 2r (or larger), the operator KL can be recovered
 let from the matrix Φ2r+2d,2r , whence the formula
c0 , . . . , ck−1 be k positive integers, and let d = i ci . KL
Given k mutually distinct points α0 , . . . , αk−1 in K and
Φ2r+2d,2r = Φ2r+2d,2r+d Φ2r+d,2r (1)
a polynomial P ∈ K[x]d , one can compute the vector KL K L

of d values yields a way to multiply K and L. For the complexity


H = 
(P (α0 ), P (α0 ), . . . , P(c0 −1)
(α0 ), . . . . . . , analysis, we thus have to consider the three steps:
P (αk−1 ), P  (αk−1 ), . . . , P (ck−1 −1) (αk−1 )) 1) (Evaluation) Computation of Φ2r+2d,2r+d
K and of
Φ2r+d,2r
L from K and L.
in O(M(d) log k) = Õ(d) arithmetic operations in K. 2) (Inner multiplication) Computation of the matrix
Conversely, P is uniquely determined by H, and its coef- product (1).
ficients can be recovered from H in O(M(d) log k) = Õ(d) 3) (Interpolation) Recovery of KL from Φ2r+2d,2r
KL .
arithmetic operations in K. In [16], [6], this multiplication method was applied with
III. The new algorithm in the case r  d success to the case when d = r. In this “square case”,
the following result was proved in [6, §4.2].
A. Multiplication by evaluation and interpolation
Lemma 1: Let L ∈ K[x, ∂]d,d . Then
Most fast algorithms for multiplying two polynomials
1) We may compute Φ2d,d as a function of L in time
P, Q ∈ K[x]d are based on the evaluation-interpolation L
O(d M(d));
strategy. The idea is to pick 2d − 1 distinct points
2) We may recover L from Φ2d,d in time O(d M(d)).
α0 , . . . , α2d−2 in K, and to perform the following three L

steps: B. Evaluation–interpolation at exponential polynomials


1) (Evaluation) Evaluate P and Q at α0 , . . . , α2d−2 .
Assume now that r  d. Then a straightforward ap-
2) (Inner multiplication) Compute the 2d − 1 values
plication of the above evaluation-interpolation strategy
(P Q)(αi ) = P (αi )Q(αi ) for i < 2d − 1.
yields an algorithm of sub-optimal complexity. Indeed,
3) (Interpolation) Recover the product P Q from the 2r+2d,2r
the matrix ΦKL contains a lot of redundant infor-
values (P Q)(α0 ), . . . , (P Q)(α2d−2 ).
mation and, since its mere total number of elements
The inner multiplication step requires only O(d) oper- exceeds r2 , one cannot expect a direct multiplication
ations. Consequently, if both the evaluation and inter- algorithm of quasi-optimal complexity Õ(dω−1 r).
polation steps can be performed fast, then we obtain a In order to maintain quasi-optimal complexity in this
fast algorithm for multiplying P and Q. For instance, if case as well, the idea is to evaluate at so called exponen-
K contains a 2p -th primitive root of unity with 2p−1  tial polynomials instead of ordinary polynomials. More
2d − 1 < 2p , then both evaluation and interpolation can specifically, given L ∈ K[x, ∂]d,r and α ∈ K, we will use
be performed in time O(d log d) using the Fast Fourier the fact that L also operates nicely on the vector space
Transform [11]. K[x]eαx . Moreover, for any P ∈ K[x], we have
For a linear differential operator L ∈ K[x, ∂]d,r it is
natural to consider evaluations at powers of x instead L(P eαx ) = Lα (P )eαx ,

526
where to a Hermite evaluation at the points αi , with multi-
plicity ci = d at each point αi . By Theorem 2, this
Lα = Li (x)(∂ + α)i
computation can be performed in time O(M(pd) log p) =
i
O(M(r) log r). Furthermore, Hermite interpolation al-
is theoperator obtained by substituting ∂ + α for ∂ in lows us to recover L from L∗α0 , . . . , L∗αp−1 with the
L = i Li (x)∂ i . Indeed, this is a consequence of the fact same time complexity O(M(r) log r).
that, by Leibniz’s rule: Now let L ∈ K[x, ∂]d,r and consider the expansion of
⎛ ⎞

i L in x
∂ i (P eαx ) = ⎝ αj ∂ i−j P ⎠ eαx = (∂ + α)i (P )eαx . L(x, ∂) = L0 (∂) + · · · + xd−1 Ld−1 (∂).
j
ji
For each i, one Hermite evaluation of Li allows us to
Now let p = r/d and let α0 , . . . , αp−1 be p pairwise
compute the L∗αj ,i with j < p in time O(M(r) log r).
distinct points in K. For each integer k  1, we define
The operators L∗αj with j < p can therefore be
the vector space
computed in time O(d M(r) log r). By Lemma 1, we
Vk = K[x]k eα0 x ⊕ · · · ⊕ K[x]k eαp−1 x need O(r M(d)) = O(d M(r)) additional operations in
[2d,d] [2d,d]
order to obtain ΦL . Similarly, given ΦL , Lemma 1
with canonical basis ∗
allows us to recover the operators Lαj with j < p in
(eα0 x , . . . , xk−1 eα0 x , . . . . . ., eαp−1 x , . . . , xk−1 eαp−1 x ). time O(d M(r)). Using d Hermite interpolations, we also
recover the coefficients Li of L in time O(d M(r) log r).
Then we may regard L as an operator from Vk into Vk+d
[k+d,k]
and we will denote by ΦL the matrix of this operator Theorem 3: Assume r  d and let K, L ∈ K[x, ∂]d,r .
with respect to the canonical bases. By what precedes, Then the product KL can be computed in time
this matrix is block diagonal, with p blocks of size d: O(dω−1 r + d M(r) log r).
⎛ k+d,k ⎞
ΦLα Proof: Considering K and L as operators in
⎜ 0
⎟ K[x, ∂]3d,3r , Lemma 2 implies that the computation of
[k+d,k] ⎜ .. ⎟
ΦL = ⎜ . ⎟. [4d,3d]
ΦK and ΦL
[3d,2d]
as a function of K and L can be
⎝ ⎠
Φk+d,k done in time O(d M(r) log r). The multiplication
Lα p−1
[4d,2d] [4d,3d] [3d,2d]
Let us now show that the operator L is uniquely deter- ΦKL = ΦK ΦL
[2d,d]
mined by the matrix ΦL , and that this gives rise to can be done in time O(dω p) = O(dω−1 r). Lemma 2
an efficient algorithm for multiplying two operators in [4d,2d]
finally implies that we may recover KL from ΦKL in
K[x, ∂]d,r . time O(d M(r) log r).
Lemma 2: Let L ∈ K[x, ∂]d,r with r  d. Then
[2d,d] IV. The new algorithm in the case d > r
1) We may compute ΦL as a function of L in time
O(d M(r) log r); Any differential operator L ∈ K[x, ∂]d,r can be written
[2d,d] in a unique form
2) We may recover L from the matrix ΦL in time
O(d M(r) log r).
 L= Li,j xj ∂ i , for some scalars Li,j ∈ K.
Proof: For any operator L = i<d, j<r Li,j xi ∂ j in i<r,j<d
K[x, ∂]d,r , we define its truncation L∗ at order O(∂ d ) by
This representation, with x on the left and ∂ on the
L∗ = Li,j xi ∂ j . right, is called the canonical form of L.
i,j<d Let ϕ : K[x, ∂] → K[x, ∂] denote the map defined by
⎛ ⎞
Since L − L∗ vanishes on K[x]d , we notice that Φ2d,d =
ϕ⎝ Li,j xj ∂ i ⎠ =
L
Φ2d,d
∗ . Li,j ∂ j (−x)i .
L
If L ∈ K[∂]r , then L∗ can be regarded as the power i<r,j<d i<r,j<d

series expansion of L at ∂ = 0 and order d. More In other words, ϕ is the unique K-algebra automorphism
generally, for any i ∈ {0, . . . , p − 1}, the operator of K[x, ∂] that keeps the elements of K fixed, and is
L∗αi (∂) = L(∂ + αi )∗ coincides with the Taylor series defined on the generators of K[x, ∂] by ϕ(x) = ∂ and
expansion at ∂ = αi and order d: ϕ(∂) = −x. We will call ϕ the reflection morphism of
K[x, ∂]. The map ϕ enjoys the nice property that it sends
L∗αi (∂) = L(αi ) + L (αi )∂ +· · ·+ (d−1)!
1
L(d−1) (αi )∂ d−1 .
K[x, ∂]d,r onto K[x, ∂]r,d . In particular, to an operator
In other words, the computation of the truncated oper- whose degree is higher than its order, ϕ associates a
ators L∗α0 , . . . , L∗αp−1 as a function of L corresponds “mirror operator” whose order is higher than its degree.

527
A. Main idea of the algorithm in the case d > r where we use the convention that pi,j = 0 as soon as
If d > r, then the reflection morphism ϕ is the key i  r or j  d.
to our fast multiplication algorithm for operators in Proof: Leibniz’s differentiation rule implies the com-
K[x, ∂]d,r , since it allows us to reduce this case to the mutation rule
previous case when r  d. More precisely, given K, L in i j

jx j xi−k j−k
K[x, ∂]d,r with d > r, the main steps of the algorithm ∂ = ∂ .
i! k (i − k)!
are: k=0
(S1) compute the canonical forms of ϕ(K) and ϕ(L), Together with the hypothesis, this implies the equality
(S2) compute the product M = ϕ(K)ϕ(L) of opera-
xi j xi
tors ϕ(K) ∈ K[x, ∂]r,d and ϕ(L) ∈ K[x, ∂]r,d , using (i! qi,j ) ∂ = (i! pi,j )∂ j
the algorithm described in the previous section, i,j
i! i,j
i!
⎛ ⎞
and

i−k
⎝ (i! pi,j ) j x
(S3) compute and return the (canonical form of the) = ∂ j−k ⎠ .
operator KL = ϕ−1 (M ). i,j
k (i − k)!
k0
Since d > r, step (S2) can be performed in complexity
Õ(rω−1 d) using the results of Section III. In the next We conclude by extraction of coefficients.
subsection, we will prove that both steps (S1) and (S3) Theorem 4: Let L ∈ K[x, ∂]d,r . Then we may compute
can be performed in Õ(rd) operations in K. This will ϕ(L) and ϕ−1 (L) using O(min(d M(r), r M(d))) = Õ(rd)
enable us to conclude the proof of Theorem 1. operations in K.
Proof:
 We first deal with the case r  d. If
B. Quasi-optimal computation of reflections L = i<r, j<d pi,j xj ∂ i , then by the first equality of
We now show that the reflection and the inverse re- Lemma 3, the reflection ϕ(L) is equal to
flection of a differential operator can be computed quasi-
optimally. The idea is that performing reflections can
be interpreted in terms of Taylor shifts for polynomials, ϕ(L) = pi,j ∂ j (−x)i = qi,j (−x)j ∂ i ,
which can be computed in quasi-linear time using the i<r,j<d i<r,j<d

algorithm from [1]. where


A first observation is that the composition ϕ ◦ ϕ is
j +

equal to the involution ψ : K[x, ∂] → K[x, ∂] defined by i! qi,j = (i +
)! pi+,j+ . (2)
⎛ ⎞ 0
j

ψ⎝ Li,j xj ∂ i ⎠ = (−1)i+j Li,j xj ∂ i . For anyfixed k with 1 − r  k  d − 1, let us introduce
i<r,j<d i<r,j<d Gk = i i!qi,i+k xi+k and Fk = i i!pi,i+k xi+k . These
As a direct consequence of this fact, it follows that polynomials belong to K[x]d , since pi,j = qi,j = 0 for
the map ϕ−1 is equal to ϕ ◦ ψ. Since ψ(L) is already j  d. If k  0, then Equation (2) translates into
in canonical form, computing ψ(L) only consists of Gk (x) = Fk (x + 1).
sign changes, which can be done in linear time O(dr).
Therefore, computing the inverse reflection ϕ−1 (L) can Indeed, Equation (2) with j = i + k implies that Gk (x)
be performed within the same cost as computing the is equal to
direct reflection ϕ(L), up to a linear overhead O(rd).
i + k +

In the remainder of this section, we focus on the fast (i +
)! pi+,i+k+ xi+k
computation of direct reflections. The key observation i+k
i,


is encapsulated in the next lemma. Here, and in what j+k s
follows, we use the convention that the entries of a = j!pj,j+k x = Fk (x + 1).
j,s
s
matrix corresponding to indices beyond the matrix sizes
are all zero. Similarly, if k > 0, then the coefficients of xi in Gk (x)
Lemma 3: Assume that (pi,j ) and (qi,j ) are two ma- and Fk (x + 1) still coincide for all i  k. In particular,
trices in Kr×d such that we may compute G1−r , . . . , Gd−1 from F1−r , . . . , Fd−1

qi,j xi ∂ j = pi,j ∂ j xi . by means of d + r  2r Taylor shifts of polynomials in
i,j i,j K[x]d . Using the fast algorithm for Taylor shift in [1],
this can be done in time O(r M(d)).
Then

j + k Once the coefficients of the Gk ’s are available, the
i! qi,j = (i + k)! pi+k,j+k , computation of the coefficients of ϕ(L) requires O(dr)
k additional operations.
k0

528
If d > r, then we notice that the equality (2) is [6] A. Bostan, F. Chyzak, and N. Le Roux, “Products of
equivalent to ordinary differential operators by evaluation and inter-
polation,” in Proceedings of ISSAC’08. New York:

i +
ACM, 2008, pp. 23–30.
j! qi,j = (j +
)! pi+,j+ ,
i [7] E. Brassinne, “Analogie des équations différentielles
0
linéaires à coefficients variables, avec les équations al-
as can be seen by expanding  the binomial coeffi- gébriques,” in Note III du Tome 2 du Cours d’analyse
cients.
 Redefining G k := i i!qi+k,i xi+k and Fk := de Ch. Sturm, École polytechnique, 2ème édition, 1864,
i+k pp. 331–347.
i i!pi+k,i x , similar arguments as above show that
ϕ(P ) can be computed using O(d M(r)) operations in K. [8] P. Bürgisser, M. Clausen, and M. A. Shokrollahi, Al-
By what has been said at the beginning of this gebraic Complexity Theory, ser. Grundlehren der Math-
section, we finally conclude that the inverse reflection ematischen Wissenschaften [Fundamental Principles of
ϕ−1 (L) = ϕ(ψ(L)) can be computed for the same cost Mathematical Sciences]. Berlin: Springer-Verlag, 1997,
as the direct reflection ϕ(L). vol. 315, with the collaboration of Thomas Lickteig.

[9] D. G. Cantor and E. Kaltofen, “On fast multiplication


C. Proof of Theorem 1 in the case d > r of polynomials over arbitrary algebras,” Acta Informat.,
We will prove a slightly better result: vol. 28, no. 7, pp. 693–701, 1991.
Theorem 5: Assume d > r and K, L ∈ K[x, ∂]d,r . [10] F. Y. Chin, “A generalized asymptotic upper bound on
Then the product KL can be computed using fast polynomial evaluation and interpolation,” SIAM J.
O(rω−1 d + r M(d) log d) operations in K. Comput., vol. 5, no. 4, pp. 682–690, 1976.
Proof: Assume that K and L are two operators in [11] J. W. Cooley and J. W. Tukey, “An algorithm for the
K[x, ∂]d,r with d > r. Then ϕ(K) and ϕ(L) belong to machine calculation of complex Fourier series,” Math.
K[x, ∂]r,d , and their canonical forms can be computed in Comp., vol. 19, pp. 297–301, 1965.
O(dM(r)) operations by Theorem 4. Using the algorithm
from Section III, we may compute M = ϕ(K)ϕ(L) [12] D. Coppersmith and S. Winograd, “Matrix multiplica-
tion via arithmetic progressions,” J. Symbolic Comput.,
in O(r ω−1 d + r M(d) log d) operations. Finally, KL = vol. 9, no. 3, pp. 251–280, Mar. 1990.
ϕ−1 (M ) can be computed in O(r M(d)) operations by
Theorem 4. We conclude by adding up the costs of these [13] S. S. Demidov, “On the history of the theory of linear
three steps. differential equations,” Arch. Hist. Exact Sci., vol. 28,
no. 4, pp. 369–387, 1983.
Acknowledgment [14] J. von zur Gathen and J. Gerhard, Modern Computer
Algebra, 2nd ed. Cambridge University Press, 2003.
The authors would like to thank the three referees for
their useful remarks. [15] D. Y. Grigor ev, “Complexity of factoring and calculat-
ing the GCD of linear ordinary differential operators,”
References J. Symbolic Comput., vol. 10, no. 1, pp. 7–37, 1990.

[1] A. V. Aho, K. Steiglitz, and J. D. Ullman, “Evaluating [16] J. van der Hoeven, “FFT-like multiplication of linear
polynomials at fixed sets of points,” SIAM J. Comput., differential operators,” J. Symbolic Comput., vol. 33,
vol. 4, no. 4, pp. 533–539, 1975. no. 1, pp. 123–127, 2002.

[2] A. V. Aho, J. E. Hopcroft, and J. D. Ullman, The design [17] ——, “On the complexity of skew arith-
and analysis of computer algorithms. Addison-Wesley metic,” Tech. Rep., 2011, HAL 00557750,
Publishing Co., 1974. http://hal.archives-ouvertes.fr/hal-00557750.

[18] G. Libri, “Mémoire sur la résolution des équations


[3] A. Benoit, A. Bostan, and J. van der Hoeven, “Fast
algébriques dont les racines ont entre elles un rap-
multiplication of skew polynomials,” in preparation.
port donné, et sur l’intégration des équations différen-
tielles linéaires dont les intégrales particulières peuvent
[4] D. Bini and V. Y. Pan, Polynomial and matrix compu- s’exprimer les unes par les autres,” J. Reine Angew.
tations. Vol. 1 – Fundamental algorithms, ser. Progress Math., vol. 10, pp. 167–194, 1833.
in Theoretical Computer Science. Boston, MA:
Birkhäuser Boston Inc., 1994. [19] R. T. Moenck, “Another polynomial homomorphism,”
Acta Informat., vol. 6, no. 2, pp. 153–169, 1976.
[5] A. Bostan, F. Chyzak, Z. Li, and B. Salvy, “Fast
computation of common left multiples of linear ordi- [20] V. Olshevsky and A. Shokrollahi, “Matrix-vector prod-
nary differential operators,” in Proceedings of ISSAC’12. uct for confluent Cauchy-like matrices with application
New York: ACM, 2012, pp. 99–106, preliminary version to confluent rational interpolation,” in Proceedings of
available at http://arxiv.org/abs/1205.0879. STOC’00. New York: ACM, 2000, pp. 573–581.

529
[21] O. Ore, “Formale Theorie der linearen Differentialgle- [25] A. Schönhage and V. Strassen, “Schnelle Multiplikation
ichungen. (Erster Teil),” J. Reine Angew. Math., vol. großer Zahlen,” Computing, vol. 7, pp. 281–292, 1971.
167, pp. 221–234, 1932.
[26] A. Stothers, “On the complexity of matrix multiplica-
[22] ——, “Theory of non-commutative polynomials,” Ann. tion,” Ph.D. dissertation, University of Edinburgh, 2010.
of Math. (2), vol. 34, no. 3, pp. 480–508, 1933.
[27] V. Strassen, “Gaussian elimination is not optimal,” Nu-
[23] V. Pan, How to multiply matrices faster, ser. Lecture mer. Math., vol. 13, pp. 354–356, 1969.
Notes in Computer Science. Berlin: Springer-Verlag,
1984, vol. 179. [28] V. Vassilevska Williams, “Multiplying matrices faster
than Coppersmith-Winograd,” in Proceedings of
[24] V. Y. Pan, Structured matrices and polynomials – Unified STOC’12. New York: ACM, 2012, pp. 887–898.
Superfast Algorithms. Boston, MA: Birkhäuser Boston
Inc., 2001.

530

You might also like