• Aucun résultat trouvé

Composable core-sets for determinant maximization: A simple near-optimal algorithm

N/A
N/A
Protected

Academic year: 2021

Partager "Composable core-sets for determinant maximization: A simple near-optimal algorithm"

Copied!
19
0
0

Texte intégral

Loading

Figure

Figure 2: Average ratio of the run time of Local Search over Greedy as a function of k.
Figure 5: Average improvement of Local Search over Greedy as a function of k, in the identical algorithms setting.

Références

Documents relatifs

This paper shows that, in complex geometry swirled combustors, such modes can appear for both cold and reacting conditions but that they have different sources: while the cold

If we use a basic clustering algorithm based on the minimization of the empirical distortion (such as the k-means algorithm) when we deal with noisy data, the expected criterion

The FrameSense algorithm, as well as systematic sampling combined with Newton’s method algorithm, the proposed Frank-Wolfe algorithm and weighted least-squares, have

We thus tune this solution, reducing inter-core communication and still allowing good pipelining (see Figure 12). The number of semaphores is then reduced to approximately 100.

In the expectation, or E-step, the missing data are estimated given the observed data and current estimate of the model parameters... achieved using the conditional

Based on the solution of this initial problem, we can further build an algorithm for finding test data set that provides coverage of the most complex branches (as many as

On the other hand, when dealing with capacities, it makes sense to study the k-additive monotone core, that consists in the set of capacities dominating the capacity; it can be

ML estimation in the presence of a nuisance vector Using the Expectation Max- imization (EM) algorithm Or the Bayesian Expectation Maximization (BEM) algorithm... ML estimation in