• Aucun résultat trouvé

On the complexity of computing certain resultants

N/A
N/A
Protected

Academic year: 2022

Partager "On the complexity of computing certain resultants"

Copied!
3
0
0

Texte intégral

(1)

ACM Communications in Computer Algebra, TBA TBA

On the complexity of computing certain resultants

Romain Lebreton

LIX, ´ Ecole polytechnique, France lebreton@lix.polytechnique.fr

Esmaeil Mehrabi and ´ Eric Schost

Computer Science Department, Western University, Canada emehrab@uwo.ca, eschost@uwo.ca

Computing resultants is a fundamental algorithmic question, at the heart of higher-level algorithms for solving systems of equations, computational topology, etc. However, in many situations, the best known algorithms are still sub-optimal. The following table summarizes the best results known to us (from [3]), using soft-Oh notation to omit logarithmic factors. In all cases, we assume that f, g have coefficients in a field k, and that their partial degrees in all variables is at most d. The partial degree in all remaining variables of their resultant r = res(f, g, x1) is then at most 2d2. In this note, the cost of an algorithm is the number of arithmetic operations in k it performs.

f, g are in . . . k[x1] k[t, x1] k[t, x0, x1]

number of terms in f, g Θ(d) Θ(d2) Θ(d3)

number of terms in r= res(f, g, x1) 1 Θ(d2) Θ(d4) cost of computing r (best known bound) O˜(d) O˜(d3) O˜(d5)

optimal? yes, up to log factors no no

In the last cases, one can replace the ringk[t] byZand consider the bit-complexity of computing resultants of polynomials in Z[x1] or Z[x0, x1]. We do not give details, but most results carry over, mutatis mutandis. One can also consider polynomials inZ[t, x0, x1], etc; we do not discuss this.

Main result. Our contribution is on the third case, withf, g ink[t, x0, x1]. We make the following assumptions (which hold for generic f, g):

A1. The reduced Groebner basis of the idealhf, giink(t)[x0, x1] for the lexicographic orderx1 > x0 has the form hR(x0), x1−S(x0)i,with R, S in k(t)[x0] andR monic.

A2. All solutions of the system f =g = 0 ink(t) have multiplicity 1.

Our algorithm uses matrix multiplication; we let 2 < ω ≤ 3 be such that one can multiply n×n matrices over k in nω operations. The best known result [5] is ω'2.37.

Theorem 1 Let f, g be ink[t, x0, x1], with degree at mostd in all variables and that satisfyA1,A2. Suppose that k has cardinality at least 12d4. There exists a probabilistic algorithm that computes r= res(f, g, x1) using O(dω+72 ) operations in k and success probability at least 1/2.

Since we have 2< ω≤3, the exponentρin our running time satisfies 4.5< ρ≤5. This improves on the best previous results, getting us closer to an optimal O˜(d4). Even under assumptions A1,A2, and allowing probabilistic algorithms, we are not aware of any previous improvements overO˜(d5).

1

(2)

Title of your paper TBA

Sketch of our algorithm. The polynomial R introduced in A1 and the resultant r of f and g are related by the equality R =r/LeadingCoefficient(r, x0). We describe how to compute R, since finding the proportionality factor that gives r is straightforward.

The algorithm uses Newton / Hensel lifting techniques. We choose a random expansion point τ for t in k. This is the source of the probabilistic aspect of the algorithm: we expect that no denominator inRorSvanishes atτ, and that the solutions of the systemf(τ, x0, x1) =g(τ, x0, x1) = 0 ink still have multiplicity 1; the analysis in [4, Prop. 3] and our assumption on the cardinality of k show that at least half the points in k are “lucky” for this random choice. Below, we take τ = 0 for simplicity; then, by assumption, forκ ≥1, Rκ =Rmodtκ and Sκ =S modtκ are well-defined;

they lie inAκ[x0], withAκ =k[t]/tκ.

We first computeR1 =Rmodtand S1 =S modt, using Reischert’s algorithm in k[x0, x1]; this costs O˜(d3). Then, we compute Rκ and Sκ for some κ ≥ 4d2 using lifting techniques: the suc- cessive lifting steps compute (R2, S2),(R4, S4), . . . ,(R2`, S2`), . . .. The assumption that the system f(τ, x0, x1) = g(τ, x0, x1) = 0 has simple roots makes this step well-defined; we analyze its cost below. Finally, we getR by applying rational reconstruction to all coefficients ofRκ in timeO˜(d4).

The key subroutine. The above process is hardly new: the references [2, 4] give details on such lifting algorithms, in more general contexts; however, as explained now, a direct application of these results performs poorly in our context.

Given (Rκ, Sκ), the algorithm of [2, 4] computes (R, S) as follows. LetBκ =A[x0, x1]/hRκ, x1− Sκi=k[t, x0, x1]/ht, Rκ, x1−Sκi. First, compute the normal form of (f, g), and of their Jacobian matrix J, in Bκ; then, deduce the vector

δR δS

=

R0κ 0

−Sκ0 1

J−1 f

g

∈Bk2×1.

Taking canonical preimages of δR and δS in A[x0], we have R = RκR and S = SκS. The bottleneck is the computation of the normal form of f, g and J: the algorithm in [2, 4] does O(d) operations in Bκ, for a total of O˜(κd3) operations in k. Summing over all lifting steps, with κ= 1,2,4,8, . . . up to about κ'd2 leads to the boundO˜(d5), which is no better than Reischert’s algorithm.

We now sketch how to compute e.g. the normal form of f inBk more efficiently, using a baby- steps / giant-steps approach inspired by Brent and Kung’s algorithm [1].

1. Seeing f as a polynomial of degree d in x1, with coefficients fi ∈ A[x0] of degree d in x0, build the √

d+ 1×√

d+ 1 matrix M1 = (f(d+1−i)d+1+j−1) with entries in A[x0].

2. Compute σ0 =Sκ0, σ1 =Sκ1,· · · , σd+1−1 =S

d+1−1

κ inBk (baby steps).

3. Cut allσi into slices, writingσi =Pd−1

j=0σi,jxdj0 , withσi,j ∈A[x0] of degree less than dinx0. Build the √

d+ 1×d matrix M2 = (σi,j) and compute M = (mij) =M1M2.

4. Using the mi,j, reconstruct f modhRκ, x1−Sκi using Horner’s scheme (giant steps).

As in Brent and Kung’s algorithm, the dominant cost is matrix multiplication (Step 3). We do matrix multiplication in sizes √

d+ 1×√

d+ 1 and √

d+ 1×d, with entries in k[t, x0], of degrees at most 2κ in t and d in x0. The cost is thus O˜(κdω+32 ) operations in k.

Summing over all lifting steps, using the fact that we take κ = 1,2,4,8, . . . (powers of 2) until approximately d2, the total cost is O˜(dω+72 ), as claimed.

2

(3)

Author’s Name

References

[1] R. P. Brent and H. T. Kung. Fast algorithms for manipulating formal power series. Journal of the ACM, 25(4):581–595, 1978.

[2] M. Giusti, G. Lecerf, and B. Salvy. A Gr¨obner free alternative for polynomial system solving.

Journal of Complexity, 17(1):154–211, 2001.

[3] D. Reischert. Asymptotically fast computation of subresultants. In ISSAC’97, pages 233–240.

ACM, 1997.

[4] ´E. Schost. Computing parametric geometric resolutions. Appl. Algebra Engrg. Comm. Comput., 13(5):349–393, 2003.

[5] V. Vassilevska Williams. Breaking the Coppersmith-Winograd barrier. 2011.

3

Références

Documents relatifs

In Chap- ter 8, we introduce complexity bounds for solving sparse polynomial systems using our algorithm to compute Gr¨obner basis over semigroup algebras.. We do not discuss how

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

Using our algorithm, we were able to compute the canonical bases of large weight sub- spaces of different Fock spaces (see examples in Section 4).. These calculations helped us

d’abord, elle est une image de la barbarie humaine sois par une force physique ,sois par les expréssionshonteuse,ce qui provoque la dinité humaine .ensuite ,elle

Via the fast conversion algorithms in Section 2, the preceding lemma enables us to reduce the computation of the composed sum of characteristic p &gt; 0 to a single multiplication

Translating the above bound, the complexity of the Montes approach is at best in O(D e 5 ) but only for computing a local integral basis at one singularity, while the algorithm

Diagnosis of the anonymous wind turbine gearbox: (a) frequency modulation index obtained in AStrion by demodulating the side-band series around the second harmonic of GMF

We have proved that if we are using the model of interval arithmetic, then the subresultant pseudo-remainder sequence algorithm will output Res(A, B) at precision O(π N − N int )