id
stringlengths 9
16
| submitter
stringlengths 4
52
⌀ | authors
stringlengths 4
937
| title
stringlengths 7
243
| comments
stringlengths 1
472
⌀ | journal-ref
stringlengths 4
244
⌀ | doi
stringlengths 14
55
⌀ | report-no
stringlengths 3
125
⌀ | categories
stringlengths 5
97
| license
stringclasses 9
values | abstract
stringlengths 33
2.95k
| versions
list | update_date
timestamp[s] | authors_parsed
sequence |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0802.3448 | Haim Kaplan | Edith Cohen and Haim Kaplan | Sketch-Based Estimation of Subpopulation-Weight | null | null | null | null | cs.DB cs.DS cs.NI cs.PF | null | Summaries of massive data sets support approximate query processing over the
original data. A basic aggregate over a set of records is the weight of
subpopulations specified as a predicate over records' attributes. Bottom-k
sketches are a powerful summarization format of weighted items that includes
priority sampling and the classic weighted sampling without replacement. They
can be computed efficiently for many representations of the data including
distributed databases and data streams.
We derive novel unbiased estimators and efficient confidence bounds for
subpopulation weight. Our estimators and bounds are tailored by distinguishing
between applications (such as data streams) where the total weight of the
sketched set can be computed by the summarization algorithm without a
significant use of additional resources, and applications (such as sketches of
network neighborhoods) where this is not the case.
Our rigorous derivations are based on clever applications of the
Horvitz-Thompson estimator, and are complemented by efficient computational
methods. We demonstrate their benefit on a wide range of Pareto distributions.
| [
{
"version": "v1",
"created": "Sat, 23 Feb 2008 15:25:04 GMT"
}
] | 2008-02-26T00:00:00 | [
[
"Cohen",
"Edith",
""
],
[
"Kaplan",
"Haim",
""
]
] |
0802.3881 | Jorge Sousa Pinto | Jos\'e Bacelar Almeida, Jorge Sousa Pinto | Deriving Sorting Algorithms | Technical Report | null | null | DI-PURe-06.04.01 | cs.DS cs.LO | null | This paper proposes new derivations of three well-known sorting algorithms,
in their functional formulation. The approach we use is based on three main
ingredients: first, the algorithms are derived from a simpler algorithm, i.e.
the specification is already a solution to the problem (in this sense our
derivations are program transformations). Secondly, a mixture of inductive and
coinductive arguments are used in a uniform, algebraic style in our reasoning.
Finally, the approach uses structural invariants so as to strengthen the
equational reasoning with logical arguments that cannot be captured in the
algebraic framework.
| [
{
"version": "v1",
"created": "Tue, 26 Feb 2008 19:47:57 GMT"
}
] | 2008-02-27T00:00:00 | [
[
"Almeida",
"José Bacelar",
""
],
[
"Pinto",
"Jorge Sousa",
""
]
] |
0802.4040 | Stephan Mertens | Stefan Boettcher, Stephan Mertens | Analysis of the Karmarkar-Karp Differencing Algorithm | 9 pages, 8 figures; minor changes | European Physics Journal B 65, 131-140 (2008) | 10.1140/epjb/e2008-00320-9 | null | cs.NA cond-mat.dis-nn cs.DM cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Karmarkar-Karp differencing algorithm is the best known polynomial time
heuristic for the number partitioning problem, fundamental in both theoretical
computer science and statistical physics. We analyze the performance of the
differencing algorithm on random instances by mapping it to a nonlinear rate
equation. Our analysis reveals strong finite size effects that explain why the
precise asymptotics of the differencing solution is hard to establish by
simulations. The asymptotic series emerging from the rate equation satisfies
all known bounds on the Karmarkar-Karp algorithm and projects a scaling
$n^{-c\ln n}$, where $c=1/(2\ln2)=0.7213...$. Our calculations reveal subtle
relations between the algorithm and Fibonacci-like sequences, and we establish
an explicit identity to that effect.
| [
{
"version": "v1",
"created": "Wed, 27 Feb 2008 17:24:07 GMT"
},
{
"version": "v2",
"created": "Fri, 3 Oct 2008 09:48:52 GMT"
}
] | 2008-10-03T00:00:00 | [
[
"Boettcher",
"Stefan",
""
],
[
"Mertens",
"Stephan",
""
]
] |
0802.4244 | Dimitris Papamichail | Christos Tryfonas, Dimitris Papamichail, Andrew Mehler, Steven Skiena | Call Admission Control Algorithm for pre-stored VBR video streams | 12 pages, 9 figures, includes appendix | null | null | null | cs.NI cs.DS | http://creativecommons.org/licenses/by-nc-sa/3.0/ | We examine the problem of accepting a new request for a pre-stored VBR video
stream that has been smoothed using any of the smoothing algorithms found in
the literature. The output of these algorithms is a piecewise constant-rate
schedule for a Variable Bit-Rate (VBR) stream. The schedule guarantees that the
decoder buffer does not overflow or underflow. The problem addressed in this
paper is the determination of the minimal time displacement of each new
requested VBR stream so that it can be accomodated by the network and/or the
video server without overbooking the committed traffic. We prove that this
call-admission control problem for multiple requested VBR streams is
NP-complete and inapproximable within a constant factor, by reducing it from
the VERTEX COLOR problem. We also present a deterministic morphology-sensitive
algorithm that calculates the minimal time displacement of a VBR stream
request. The complexity of the proposed algorithm make it suitable for
real-time determination of the time displacement parameter during the call
admission phase.
| [
{
"version": "v1",
"created": "Thu, 28 Feb 2008 17:45:03 GMT"
}
] | 2008-02-29T00:00:00 | [
[
"Tryfonas",
"Christos",
""
],
[
"Papamichail",
"Dimitris",
""
],
[
"Mehler",
"Andrew",
""
],
[
"Skiena",
"Steven",
""
]
] |
0802.4325 | Mirela Damian | Mirela Damian | A Simple Yao-Yao-Based Spanner of Bounded Degree | 7 pages, 5 figures | null | null | null | cs.CG cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | It is a standing open question to decide whether the Yao-Yao structure for
unit disk graphs (UDGs) is a length spanner of not. This question is highly
relevant to the topology control problem for wireless ad hoc networks. In this
paper we make progress towards resolving this question by showing that the
Yao-Yao structure is a length spanner for UDGs of bounded aspect ratio. We also
propose a new local algorithm, called Yao-Sparse-Sink, based on the Yao-Sink
method introduced by Li, Wan, Wang and Frieder, that computes a (1+e)-spanner
of bounded degree for a given UDG and for given e > 0. The Yao-Sparse-Sink
method enables an efficient local computation of sparse sink trees. Finally, we
show that all these structures for UDGs -- Yao, Yao-Yao, Yao-Sink and
Yao-Sparse-Sink -- have arbitrarily large weight.
| [
{
"version": "v1",
"created": "Fri, 29 Feb 2008 14:39:59 GMT"
},
{
"version": "v2",
"created": "Fri, 4 Apr 2008 14:40:40 GMT"
}
] | 2008-04-04T00:00:00 | [
[
"Damian",
"Mirela",
""
]
] |
0803.0248 | Emmanuelle Lebhar | Augustin Chaintreau, Pierre Fraigniaud, Emmanuelle Lebhar | Networks become navigable as nodes move and forget | 21 pages, 1 figure | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a dynamical process for network evolution, aiming at explaining
the emergence of the small world phenomenon, i.e., the statistical observation
that any pair of individuals are linked by a short chain of acquaintances
computable by a simple decentralized routing algorithm, known as greedy
routing. Previously proposed dynamical processes enabled to demonstrate
experimentally (by simulations) that the small world phenomenon can emerge from
local dynamics. However, the analysis of greedy routing using the probability
distributions arising from these dynamics is quite complex because of mutual
dependencies. In contrast, our process enables complete formal analysis. It is
based on the combination of two simple processes: a random walk process, and an
harmonic forgetting process. Both processes reflect natural behaviors of the
individuals, viewed as nodes in the network of inter-individual acquaintances.
We prove that, in k-dimensional lattices, the combination of these two
processes generates long-range links mutually independently distributed as a
k-harmonic distribution. We analyze the performances of greedy routing at the
stationary regime of our process, and prove that the expected number of steps
for routing from any source to any target in any multidimensional lattice is a
polylogarithmic function of the distance between the two nodes in the lattice.
Up to our knowledge, these results are the first formal proof that navigability
in small worlds can emerge from a dynamical process for network evolution. Our
dynamical process can find practical applications to the design of spatial
gossip and resource location protocols.
| [
{
"version": "v1",
"created": "Mon, 3 Mar 2008 14:44:08 GMT"
}
] | 2008-03-04T00:00:00 | [
[
"Chaintreau",
"Augustin",
""
],
[
"Fraigniaud",
"Pierre",
""
],
[
"Lebhar",
"Emmanuelle",
""
]
] |
0803.0473 | Edith Cohen | Edith Cohen, Nick Duffield, Haim Kaplan, Carsten Lund, and Mikkel
Thorup | Stream sampling for variance-optimal estimation of subset sums | 31 pages. An extended abstract appeared in the proceedings of the
20th ACM-SIAM Symposium on Discrete Algorithms (SODA 2009) | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | From a high volume stream of weighted items, we want to maintain a generic
sample of a certain limited size $k$ that we can later use to estimate the
total weight of arbitrary subsets. This is the classic context of on-line
reservoir sampling, thinking of the generic sample as a reservoir. We present
an efficient reservoir sampling scheme, $\varoptk$, that dominates all previous
schemes in terms of estimation quality.
$\varoptk$ provides {\em variance optimal unbiased estimation of subset
sums}. More precisely, if we have seen $n$ items of the stream, then for {\em
any} subset size $m$, our scheme based on $k$ samples minimizes the average
variance over all subsets of size $m$. In fact, the optimality is against any
off-line scheme with $k$ samples tailored for the concrete set of items seen.
In addition to optimal average variance, our scheme provides tighter worst-case
bounds on the variance of {\em particular} subsets than previously possible. It
is efficient, handling each new item of the stream in $O(\log k)$ time.
Finally, it is particularly well suited for combination of samples from
different streams in a distributed setting.
| [
{
"version": "v1",
"created": "Tue, 4 Mar 2008 15:12:24 GMT"
},
{
"version": "v2",
"created": "Mon, 15 Nov 2010 16:43:54 GMT"
}
] | 2010-11-16T00:00:00 | [
[
"Cohen",
"Edith",
""
],
[
"Duffield",
"Nick",
""
],
[
"Kaplan",
"Haim",
""
],
[
"Lund",
"Carsten",
""
],
[
"Thorup",
"Mikkel",
""
]
] |
0803.0476 | Renaud Lambiotte | Vincent D. Blondel, Jean-Loup Guillaume, Renaud Lambiotte and Etienne
Lefebvre | Fast unfolding of communities in large networks | 6 pages, 5 figures, 1 table; new version with new figures in order to
clarify our method, where we look more carefully at the role played by the
ordering of the nodes and where we compare our method with that of Wakita and
Tsurumi | J. Stat. Mech. (2008) P10008 | 10.1088/1742-5468/2008/10/P10008 | null | physics.soc-ph cond-mat.stat-mech cs.CY cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a simple method to extract the community structure of large
networks. Our method is a heuristic method that is based on modularity
optimization. It is shown to outperform all other known community detection
method in terms of computation time. Moreover, the quality of the communities
detected is very good, as measured by the so-called modularity. This is shown
first by identifying language communities in a Belgian mobile phone network of
2.6 million customers and by analyzing a web graph of 118 million nodes and
more than one billion links. The accuracy of our algorithm is also verified on
ad-hoc modular networks. .
| [
{
"version": "v1",
"created": "Tue, 4 Mar 2008 15:29:44 GMT"
},
{
"version": "v2",
"created": "Fri, 25 Jul 2008 09:52:42 GMT"
}
] | 2008-12-01T00:00:00 | [
[
"Blondel",
"Vincent D.",
""
],
[
"Guillaume",
"Jean-Loup",
""
],
[
"Lambiotte",
"Renaud",
""
],
[
"Lefebvre",
"Etienne",
""
]
] |
0803.0701 | Gregory Gutin | N Alon, F.V. Fomin, G. Gutin, M. Krivelevich and S. Saurabh | Spanning directed trees with many leaves | null | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The {\sc Directed Maximum Leaf Out-Branching} problem is to find an
out-branching (i.e. a rooted oriented spanning tree) in a given digraph with
the maximum number of leaves. In this paper, we obtain two combinatorial
results on the number of leaves in out-branchings. We show that
- every strongly connected $n$-vertex digraph $D$ with minimum in-degree at
least 3 has an out-branching with at least $(n/4)^{1/3}-1$ leaves;
- if a strongly connected digraph $D$ does not contain an out-branching with
$k$ leaves, then the pathwidth of its underlying graph UG($D$) is $O(k\log k)$.
Moreover, if the digraph is acyclic, the pathwidth is at most $4k$.
The last result implies that it can be decided in time $2^{O(k\log^2 k)}\cdot
n^{O(1)}$ whether a strongly connected digraph on $n$ vertices has an
out-branching with at least $k$ leaves. On acyclic digraphs the running time of
our algorithm is $2^{O(k\log k)}\cdot n^{O(1)}$.
| [
{
"version": "v1",
"created": "Wed, 5 Mar 2008 16:38:34 GMT"
}
] | 2008-03-06T00:00:00 | [
[
"Alon",
"N",
""
],
[
"Fomin",
"F. V.",
""
],
[
"Gutin",
"G.",
""
],
[
"Krivelevich",
"M.",
""
],
[
"Saurabh",
"S.",
""
]
] |
0803.0726 | Marie-Pierre B\'eal | Marie-Pierre B\'eal, Dominique Perrin | A quadratic algorithm for road coloring | null | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Road Coloring Theorem states that every aperiodic directed graph with
constant out-degree has a synchronized coloring. This theorem had been
conjectured during many years as the Road Coloring Problem before being settled
by A. Trahtman. Trahtman's proof leads to an algorithm that finds a
synchronized labeling with a cubic worst-case time complexity. We show a
variant of his construction with a worst-case complexity which is quadratic in
time and linear in space. We also extend the Road Coloring Theorem to the
periodic case.
| [
{
"version": "v1",
"created": "Wed, 5 Mar 2008 20:35:54 GMT"
},
{
"version": "v2",
"created": "Wed, 5 Mar 2008 21:33:23 GMT"
},
{
"version": "v3",
"created": "Mon, 7 Apr 2008 15:12:00 GMT"
},
{
"version": "v4",
"created": "Tue, 15 Apr 2008 16:32:12 GMT"
},
{
"version": "v5",
"created": "Wed, 14 May 2008 14:54:09 GMT"
},
{
"version": "v6",
"created": "Fri, 11 Jul 2008 14:21:07 GMT"
},
{
"version": "v7",
"created": "Wed, 7 Oct 2009 16:00:57 GMT"
},
{
"version": "v8",
"created": "Mon, 23 Apr 2012 16:17:48 GMT"
},
{
"version": "v9",
"created": "Thu, 30 May 2013 16:16:40 GMT"
}
] | 2013-05-31T00:00:00 | [
[
"Béal",
"Marie-Pierre",
""
],
[
"Perrin",
"Dominique",
""
]
] |
0803.0731 | Ning Chen | Ning Chen and Zhiyuan Yan | Complexity Analysis of Reed-Solomon Decoding over GF(2^m) Without Using
Syndromes | 11 pages, submitted to EURASIP Journal on Wireless Communications and
Networking | null | null | null | cs.IT cs.CC cs.DS math.IT | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | For the majority of the applications of Reed-Solomon (RS) codes, hard
decision decoding is based on syndromes. Recently, there has been renewed
interest in decoding RS codes without using syndromes. In this paper, we
investigate the complexity of syndromeless decoding for RS codes, and compare
it to that of syndrome-based decoding. Aiming to provide guidelines to
practical applications, our complexity analysis differs in several aspects from
existing asymptotic complexity analysis, which is typically based on
multiplicative fast Fourier transform (FFT) techniques and is usually in big O
notation. First, we focus on RS codes over characteristic-2 fields, over which
some multiplicative FFT techniques are not applicable. Secondly, due to
moderate block lengths of RS codes in practice, our analysis is complete since
all terms in the complexities are accounted for. Finally, in addition to fast
implementation using additive FFT techniques, we also consider direct
implementation, which is still relevant for RS codes with moderate lengths.
Comparing the complexities of both syndromeless and syndrome-based decoding
algorithms based on direct and fast implementations, we show that syndromeless
decoding algorithms have higher complexities than syndrome-based ones for high
rate RS codes regardless of the implementation. Both errors-only and
errors-and-erasures decoding are considered in this paper. We also derive
tighter bounds on the complexities of fast polynomial multiplications based on
Cantor's approach and the fast extended Euclidean algorithm.
| [
{
"version": "v1",
"created": "Wed, 5 Mar 2008 18:54:35 GMT"
},
{
"version": "v2",
"created": "Wed, 7 May 2008 21:05:41 GMT"
}
] | 2008-05-08T00:00:00 | [
[
"Chen",
"Ning",
""
],
[
"Yan",
"Zhiyuan",
""
]
] |
0803.0792 | Siddhartha Sen | Bernhard Haeupler, Siddhartha Sen, and Robert E. Tarjan | Incremental Topological Ordering and Strong Component Maintenance | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present an on-line algorithm for maintaining a topological order of a
directed acyclic graph as arcs are added, and detecting a cycle when one is
created. Our algorithm takes O(m^{1/2}) amortized time per arc, where m is the
total number of arcs. For sparse graphs, this bound improves the best previous
bound by a logarithmic factor and is tight to within a constant factor for a
natural class of algorithms that includes all the existing ones. Our main
insight is that the bidirectional search method of previous algorithms does not
require an ordered search, but can be more general. This allows us to avoid the
use of heaps (priority queues) entirely. Instead, the deterministic version of
our algorithm uses (approximate) median-finding. The randomized version of our
algorithm avoids this complication, making it very simple. We extend our
topological ordering algorithm to give the first detailed algorithm for
maintaining the strong components of a directed graph, and a topological order
of these components, as arcs are added. This extension also has an amortized
time bound of O(m^{1/2}) per arc.
| [
{
"version": "v1",
"created": "Thu, 6 Mar 2008 05:11:18 GMT"
}
] | 2008-03-07T00:00:00 | [
[
"Haeupler",
"Bernhard",
""
],
[
"Sen",
"Siddhartha",
""
],
[
"Tarjan",
"Robert E.",
""
]
] |
0803.0845 | Evain Laurent | Laurent Evain | Knapsack cryptosystems built on NP-hard instance | 20 pages | null | null | null | cs.CR cs.CC cs.DM cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We construct three public key knapsack cryptosystems. Standard knapsack
cryptosystems hide easy instances of the knapsack problem and have been broken.
The systems considered in the article face this problem: They hide a random
(possibly hard) instance of the knapsack problem. We provide both complexity
results (size of the key, time needed to encypher/decypher...) and experimental
results. Security results are given for the second cryptosystem (the fastest
one and the one with the shortest key). Probabilistic polynomial reductions
show that finding the private key is as difficult as factorizing a product of
two primes. We also consider heuristic attacks. First, the density of the
cryptosystem can be chosen arbitrarily close to one, discarding low density
attacks. Finally, we consider explicit heuristic attacks based on the LLL
algorithm and we prove that with respect to these attacks, the public key is as
secure as a random key.
| [
{
"version": "v1",
"created": "Thu, 6 Mar 2008 12:20:35 GMT"
}
] | 2008-03-17T00:00:00 | [
[
"Evain",
"Laurent",
""
]
] |
0803.0929 | Daniel A. Spielman | Daniel A. Spielman, Nikhil Srivastava | Graph Sparsification by Effective Resistances | null | null | null | null | cs.DS | http://creativecommons.org/licenses/by/3.0/ | We present a nearly-linear time algorithm that produces high-quality
sparsifiers of weighted graphs. Given as input a weighted graph $G=(V,E,w)$ and
a parameter $\epsilon>0$, we produce a weighted subgraph
$H=(V,\tilde{E},\tilde{w})$ of $G$ such that $|\tilde{E}|=O(n\log
n/\epsilon^2)$ and for all vectors $x\in\R^V$ $(1-\epsilon)\sum_{uv\in
E}(x(u)-x(v))^2w_{uv}\le \sum_{uv\in\tilde{E}}(x(u)-x(v))^2\tilde{w}_{uv} \le
(1+\epsilon)\sum_{uv\in E}(x(u)-x(v))^2w_{uv}. (*)$
This improves upon the sparsifiers constructed by Spielman and Teng, which
had $O(n\log^c n)$ edges for some large constant $c$, and upon those of
Bencz\'ur and Karger, which only satisfied (*) for $x\in\{0,1\}^V$.
A key ingredient in our algorithm is a subroutine of independent interest: a
nearly-linear time algorithm that builds a data structure from which we can
query the approximate effective resistance between any two vertices in a graph
in $O(\log n)$ time.
| [
{
"version": "v1",
"created": "Thu, 6 Mar 2008 18:03:06 GMT"
},
{
"version": "v2",
"created": "Fri, 7 Mar 2008 23:10:59 GMT"
},
{
"version": "v3",
"created": "Fri, 14 Mar 2008 19:49:32 GMT"
},
{
"version": "v4",
"created": "Wed, 18 Nov 2009 07:22:03 GMT"
}
] | 2009-11-18T00:00:00 | [
[
"Spielman",
"Daniel A.",
""
],
[
"Srivastava",
"Nikhil",
""
]
] |
0803.0954 | Michael Hahsler | Michael Hahsler, Christian Buchta, and Kurt Hornik | Selective association rule generation | null | Computational Statistics, 2007. Online First, Published: 25 July
2007 | 10.1007/s00180-007-0062-z | null | cs.DB cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Mining association rules is a popular and well researched method for
discovering interesting relations between variables in large databases. A
practical problem is that at medium to low support values often a large number
of frequent itemsets and an even larger number of association rules are found
in a database. A widely used approach is to gradually increase minimum support
and minimum confidence or to filter the found rules using increasingly strict
constraints on additional measures of interestingness until the set of rules
found is reduced to a manageable size. In this paper we describe a different
approach which is based on the idea to first define a set of ``interesting''
itemsets (e.g., by a mixture of mining and expert knowledge) and then, in a
second step to selectively generate rules for only these itemsets. The main
advantage of this approach over increasing thresholds or filtering rules is
that the number of rules found is significantly reduced while at the same time
it is not necessary to increase the support and confidence thresholds which
might lead to missing important information in the database.
| [
{
"version": "v1",
"created": "Thu, 6 Mar 2008 19:43:35 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Hahsler",
"Michael",
""
],
[
"Buchta",
"Christian",
""
],
[
"Hornik",
"Kurt",
""
]
] |
0803.0988 | Samuel Daitch | Samuel I. Daitch, Daniel A. Spielman | Faster Approximate Lossy Generalized Flow via Interior Point Algorithms | v2: bug fixes and some expanded proofs | null | null | null | cs.DS cs.NA | http://creativecommons.org/licenses/by/3.0/ | We present faster approximation algorithms for generalized network flow
problems. A generalized flow is one in which the flow out of an edge differs
from the flow into the edge by a constant factor. We limit ourselves to the
lossy case, when these factors are at most 1.
Our algorithm uses a standard interior-point algorithm to solve a linear
program formulation of the network flow problem. The system of linear equations
that arises at each step of the interior-point algorithm takes the form of a
symmetric M-matrix. We present an algorithm for solving such systems in nearly
linear time. The algorithm relies on the Spielman-Teng nearly linear time
algorithm for solving linear systems in diagonally-dominant matrices.
For a graph with m edges, our algorithm obtains an additive epsilon
approximation of the maximum generalized flow and minimum cost generalized flow
in time tildeO(m^(3/2) * log(1/epsilon)). In many parameter ranges, this
improves over previous algorithms by a factor of approximately m^(1/2). We also
obtain a similar improvement for exactly solving the standard min-cost flow
problem.
| [
{
"version": "v1",
"created": "Thu, 6 Mar 2008 21:57:53 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Apr 2008 19:02:38 GMT"
}
] | 2008-04-07T00:00:00 | [
[
"Daitch",
"Samuel I.",
""
],
[
"Spielman",
"Daniel A.",
""
]
] |
0803.1245 | George Bell | George I. Bell | The shortest game of Chinese Checkers and related problems | 22 pages, 10 figures; published version | INTEGERS: Electronic Journal of Combinatorial Number Theory 9
(2009) #G01 | null | null | math.CO cs.DM cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In 1979, David Fabian found a complete game of two-person Chinese Checkers in
30 moves (15 by each player) [Martin Gardner, Penrose Tiles to Trapdoor
Ciphers, MAA, 1997]. This solution requires that the two players cooperate to
generate a win as quickly as possible for one of them. We show, using
computational search techniques, that no shorter game is possible. We also
consider a solitaire version of Chinese Checkers where one player attempts to
move her pieces across the board in as few moves as possible. In 1971, Octave
Levenspiel found a solution in 27 moves [Ibid.]; we demonstrate that no shorter
solution exists. To show optimality, we employ a variant of A* search, as well
as bidirectional search.
| [
{
"version": "v1",
"created": "Sat, 8 Mar 2008 14:38:31 GMT"
},
{
"version": "v2",
"created": "Tue, 13 Jan 2009 18:41:21 GMT"
}
] | 2009-01-13T00:00:00 | [
[
"Bell",
"George I.",
""
]
] |
0803.1321 | Yngve Villanger | Fedor V. Fomin and Yngve Villanger | Treewidth computation and extremal combinatorics | Corrected typos | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | For a given graph G and integers b,f >= 0, let S be a subset of vertices of G
of size b+1 such that the subgraph of G induced by S is connected and S can be
separated from other vertices of G by removing f vertices. We prove that every
graph on n vertices contains at most n\binom{b+f}{b} such vertex subsets. This
result from extremal combinatorics appears to be very useful in the design of
several enumeration and exact algorithms. In particular, we use it to provide
algorithms that for a given n-vertex graph G - compute the treewidth of G in
time O(1.7549^n) by making use of exponential space and in time O(2.6151^n) and
polynomial space; - decide in time O(({\frac{2n+k+1}{3})^{k+1}\cdot kn^6}) if
the treewidth of G is at most k; - list all minimal separators of G in time
O(1.6181^n) and all potential maximal cliques of G in time O(1.7549^n). This
significantly improves previous algorithms for these problems.
| [
{
"version": "v1",
"created": "Sun, 9 Mar 2008 20:54:58 GMT"
},
{
"version": "v2",
"created": "Mon, 5 May 2008 09:34:16 GMT"
}
] | 2008-05-05T00:00:00 | [
[
"Fomin",
"Fedor V.",
""
],
[
"Villanger",
"Yngve",
""
]
] |
0803.2174 | Mirela Damian | Mirela Damian, Saurav Pandit and Sriram Pemmaraju | Local Approximation Schemes for Topology Control | 11 pages, 6 figures | Proceedings of the 25th ACM Symposium on Principles of Distributed
Computing, pages 208-218, July 2006 | null | null | cs.DS cs.CC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents a distributed algorithm on wireless ad-hoc networks that
runs in polylogarithmic number of rounds in the size of the network and
constructs a linear size, lightweight, (1+\epsilon)-spanner for any given
\epsilon > 0. A wireless network is modeled by a d-dimensional \alpha-quasi
unit ball graph (\alpha-UBG), which is a higher dimensional generalization of
the standard unit disk graph (UDG) model. The d-dimensional \alpha-UBG model
goes beyond the unrealistic ``flat world'' assumption of UDGs and also takes
into account transmission errors, fading signal strength, and physical
obstructions. The main result in the paper is this: for any fixed \epsilon > 0,
0 < \alpha \le 1, and d \ge 2, there is a distributed algorithm running in
O(\log n \log^* n) communication rounds on an n-node, d-dimensional \alpha-UBG
G that computes a (1+\epsilon)-spanner G' of G with maximum degree \Delta(G') =
O(1) and total weight w(G') = O(w(MST(G)). This result is motivated by the
topology control problem in wireless ad-hoc networks and improves on existing
topology control algorithms along several dimensions. The technical
contributions of the paper include a new, sequential, greedy algorithm with
relaxed edge ordering and lazy updating, and clustering techniques for
filtering out unnecessary edges.
| [
{
"version": "v1",
"created": "Fri, 14 Mar 2008 14:37:12 GMT"
}
] | 2008-03-17T00:00:00 | [
[
"Damian",
"Mirela",
""
],
[
"Pandit",
"Saurav",
""
],
[
"Pemmaraju",
"Sriram",
""
]
] |
0803.2615 | Olivier Laval | Olivier Laval (LIPN), Sophie Toulouse (LIPN), Anass Nagih (LITA) | Rapport de recherche sur le probl\`eme du plus court chemin contraint | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This article provides an overview of the performance and the theoretical
complexity of approximate and exact methods for various versions of the
shortest path problem. The proposed study aims to improve the resolution of a
more general covering problem within a column generation scheme in which the
shortest path problem is the sub-problem.
| [
{
"version": "v1",
"created": "Tue, 18 Mar 2008 12:37:36 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Laval",
"Olivier",
"",
"LIPN"
],
[
"Toulouse",
"Sophie",
"",
"LIPN"
],
[
"Nagih",
"Anass",
"",
"LITA"
]
] |
0803.2842 | Shai Gutner | Noga Alon, Yossi Azar, Shai Gutner | Admission Control to Minimize Rejections and Online Set Cover with
Repetitions | null | Proc. of 17th SPAA (2005), 238-244 | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study the admission control problem in general networks. Communication
requests arrive over time, and the online algorithm accepts or rejects each
request while maintaining the capacity limitations of the network. The
admission control problem has been usually analyzed as a benefit problem, where
the goal is to devise an online algorithm that accepts the maximum number of
requests possible. The problem with this objective function is that even
algorithms with optimal competitive ratios may reject almost all of the
requests, when it would have been possible to reject only a few. This could be
inappropriate for settings in which rejections are intended to be rare events.
In this paper, we consider preemptive online algorithms whose goal is to
minimize the number of rejected requests. Each request arrives together with
the path it should be routed on. We show an $O(\log^2 (mc))$-competitive
randomized algorithm for the weighted case, where $m$ is the number of edges in
the graph and $c$ is the maximum edge capacity. For the unweighted case, we
give an $O(\log m \log c)$-competitive randomized algorithm. This settles an
open question of Blum, Kalai and Kleinberg raised in \cite{BlKaKl01}. We note
that allowing preemption and handling requests with given paths are essential
for avoiding trivial lower bounds.
| [
{
"version": "v1",
"created": "Wed, 19 Mar 2008 16:53:42 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Alon",
"Noga",
""
],
[
"Azar",
"Yossi",
""
],
[
"Gutner",
"Shai",
""
]
] |
0803.3531 | Daniel Raible | Daniel Raible and Henning Fernau | A New Upper Bound for Max-2-Sat: A Graph-Theoretic Approach | null | null | null | null | cs.DS cs.CC cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In {\sc MaxSat}, we ask for an assignment which satisfies the maximum number
of clauses for a boolean formula in CNF. We present an algorithm yielding a run
time upper bound of $O^*(2^{\frac{1}{6.2158}})$ for {\sc Max-2-Sat} (each
clause contains at most 2 literals), where $K$ is the number of clauses. The
run time has been achieved by using heuristic priorities on the choice of the
variable on which we branch. The implementation of these heuristic priorities
is rather simple, though they have a significant effect on the run time. The
analysis is done using a tailored non-standard measure.
| [
{
"version": "v1",
"created": "Tue, 25 Mar 2008 11:32:22 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Apr 2008 08:36:18 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Raible",
"Daniel",
""
],
[
"Fernau",
"Henning",
""
]
] |
0803.3632 | Mikhail Nesterenko | Mikhail Nesterenko, Adnan Vora | Void Traversal for Guaranteed Delivery in Geometric Routing | null | The 2nd IEEE International Conference on Mobile Ad-hoc and Sensor
Systems (MASS 2005), Washington, DC, November, 2005 | 10.1109/MAHSS.2005.1542862 | null | cs.OS cs.DC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Geometric routing algorithms like GFG (GPSR) are lightweight, scalable
algorithms that can be used to route in resource-constrained ad hoc wireless
networks. However, such algorithms run on planar graphs only. To efficiently
construct a planar graph, they require a unit-disk graph. To make the topology
unit-disk, the maximum link length in the network has to be selected
conservatively. In practical setting this leads to the designs where the node
density is rather high. Moreover, the network diameter of a planar subgraph is
greater than the original graph, which leads to longer routes. To remedy this
problem, we propose a void traversal algorithm that works on arbitrary
geometric graphs. We describe how to use this algorithm for geometric routing
with guaranteed delivery and compare its performance with GFG.
| [
{
"version": "v1",
"created": "Tue, 25 Mar 2008 20:52:17 GMT"
}
] | 2016-11-15T00:00:00 | [
[
"Nesterenko",
"Mikhail",
""
],
[
"Vora",
"Adnan",
""
]
] |
0803.3657 | Yeow Meng Chee | Yeow Meng Chee and San Ling | Improved Lower Bounds for Constant GC-Content DNA Codes | 4 pages | IEEE Transactions on Information Theory, vol. 54, no. 1, pp.
391-394, 2008 | 10.1109/TIT.2007.911167 | null | cs.IT cs.DS math.CO math.IT q-bio.GN q-bio.QM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The design of large libraries of oligonucleotides having constant GC-content
and satisfying Hamming distance constraints between oligonucleotides and their
Watson-Crick complements is important in reducing hybridization errors in DNA
computing, DNA microarray technologies, and molecular bar coding. Various
techniques have been studied for the construction of such oligonucleotide
libraries, ranging from algorithmic constructions via stochastic local search
to theoretical constructions via coding theory. We introduce a new stochastic
local search method which yields improvements up to more than one third of the
benchmark lower bounds of Gaborit and King (2005) for n-mer oligonucleotide
libraries when n <= 14. We also found several optimal libraries by computing
maximum cliques on certain graphs.
| [
{
"version": "v1",
"created": "Wed, 26 Mar 2008 02:26:36 GMT"
}
] | 2008-03-27T00:00:00 | [
[
"Chee",
"Yeow Meng",
""
],
[
"Ling",
"San",
""
]
] |
0803.3693 | Rasmus Pagh | Martin Dietzfelbinger and Rasmus Pagh | Succinct Data Structures for Retrieval and Approximate Membership | null | null | null | null | cs.DS cs.DB cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The retrieval problem is the problem of associating data with keys in a set.
Formally, the data structure must store a function f: U ->{0,1}^r that has
specified values on the elements of a given set S, a subset of U, |S|=n, but
may have any value on elements outside S. Minimal perfect hashing makes it
possible to avoid storing the set S, but this induces a space overhead of
Theta(n) bits in addition to the nr bits needed for function values. In this
paper we show how to eliminate this overhead. Moreover, we show that for any k
query time O(k) can be achieved using space that is within a factor 1+e^{-k} of
optimal, asymptotically for large n. If we allow logarithmic evaluation time,
the additive overhead can be reduced to O(log log n) bits whp. The time to
construct the data structure is O(n), expected. A main technical ingredient is
to utilize existing tight bounds on the probability of almost square random
matrices with rows of low weight to have full row rank. In addition to direct
constructions, we point out a close connection between retrieval structures and
hash tables where keys are stored in an array and some kind of probing scheme
is used. Further, we propose a general reduction that transfers the results on
retrieval into analogous results on approximate membership, a problem
traditionally addressed using Bloom filters. Again, we show how to eliminate
the space overhead present in previously known methods, and get arbitrarily
close to the lower bound. The evaluation procedures of our data structures are
extremely simple (similar to a Bloom filter). For the results stated above we
assume free access to fully random hash functions. However, we show how to
justify this assumption using extra space o(n) to simulate full randomness on a
RAM.
| [
{
"version": "v1",
"created": "Wed, 26 Mar 2008 10:53:49 GMT"
}
] | 2008-03-27T00:00:00 | [
[
"Dietzfelbinger",
"Martin",
""
],
[
"Pagh",
"Rasmus",
""
]
] |
0803.3746 | Leonid Litinskii | Leonid B. Litinskii | Cluster Approach to the Domains Formation | 11 pages, 5 figures, PDF-file | Optical Memory & Neural Networks (Information Optics), 2007,
v.16(3) pp.144-153 | null | null | cs.NE cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | As a rule, a quadratic functional depending on a great number of binary
variables has a lot of local minima. One of approaches allowing one to find in
averaged deeper local minima is aggregation of binary variables into larger
blocks/domains. To minimize the functional one has to change the states of
aggregated variables (domains). In the present publication we discuss methods
of domains formation. It is shown that the best results are obtained when
domains are formed by variables that are strongly connected with each other.
| [
{
"version": "v1",
"created": "Wed, 26 Mar 2008 15:14:33 GMT"
}
] | 2008-03-27T00:00:00 | [
[
"Litinskii",
"Leonid B.",
""
]
] |
0803.4260 | Xin Han | Xin Han, Kazuo Iwama, Guochuan Zhang | On Two Dimensional Orthogonal Knapsack Problem | null | null | null | null | cs.DS cs.CC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we study the following knapsack problem: Given a list of
squares with profits, we are requested to pack a sublist of them into a
rectangular bin (not a unit square bin) to make profits in the bin as large as
possible. We first observe there is a Polynomial Time Approximation Scheme
(PTAS) for the problem of packing weighted squares into rectangular bins with
large resources, then apply the PTAS to the problem of packing squares with
profits into a rectangular bin and get a $\frac65+\epsilon$ approximation
algorithm.
| [
{
"version": "v1",
"created": "Sat, 29 Mar 2008 11:15:11 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Han",
"Xin",
""
],
[
"Iwama",
"Kazuo",
""
],
[
"Zhang",
"Guochuan",
""
]
] |
0803.4355 | Marko A. Rodriguez | Marko A. Rodriguez | Grammar-Based Random Walkers in Semantic Networks | First draft of manuscript originally written in November 2006 | Rodriguez, M.A., "Grammar-Based Random Walkers in Semantic
Networks", Knowledge-Based Systems, volume 21, issue 7, pages 727-739, ISSN:
0950-7051, Elsevier, October 2008 | 10.1016/j.knosys.2008.03.030 | LA-UR-06-7791 | cs.AI cs.DS | http://creativecommons.org/licenses/publicdomain/ | Semantic networks qualify the meaning of an edge relating any two vertices.
Determining which vertices are most "central" in a semantic network is
difficult because one relationship type may be deemed subjectively more
important than another. For this reason, research into semantic network metrics
has focused primarily on context-based rankings (i.e. user prescribed
contexts). Moreover, many of the current semantic network metrics rank semantic
associations (i.e. directed paths between two vertices) and not the vertices
themselves. This article presents a framework for calculating semantically
meaningful primary eigenvector-based metrics such as eigenvector centrality and
PageRank in semantic networks using a modified version of the random walker
model of Markov chain analysis. Random walkers, in the context of this article,
are constrained by a grammar, where the grammar is a user defined data
structure that determines the meaning of the final vertex ranking. The ideas in
this article are presented within the context of the Resource Description
Framework (RDF) of the Semantic Web initiative.
| [
{
"version": "v1",
"created": "Mon, 31 Mar 2008 00:13:26 GMT"
},
{
"version": "v2",
"created": "Wed, 10 Sep 2008 23:58:07 GMT"
}
] | 2008-09-11T00:00:00 | [
[
"Rodriguez",
"Marko A.",
""
]
] |
0804.0149 | Fabien Mathieu | Bruno Gaume (IRIT), Fabien Mathieu (FT R&D, INRIA Rocquencourt) | From Random Graph to Small World by Wandering | null | null | null | RR-6489 | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Numerous studies show that most known real-world complex networks share
similar properties in their connectivity and degree distribution. They are
called small worlds. This article gives a method to turn random graphs into
Small World graphs by the dint of random walks.
| [
{
"version": "v1",
"created": "Tue, 1 Apr 2008 11:59:43 GMT"
},
{
"version": "v2",
"created": "Wed, 2 Apr 2008 08:12:38 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Gaume",
"Bruno",
"",
"IRIT"
],
[
"Mathieu",
"Fabien",
"",
"FT R&D, INRIA Rocquencourt"
]
] |
0804.0277 | Marko A. Rodriguez | Marko A. Rodriguez | Mapping Semantic Networks to Undirected Networks | null | International Journal of Applied Mathematics and Computer
Sciences, volume 5, issue 1, pages 39-42, ISSN:2070-3902, LA-UR-07-5287, 2009 | null | LAUR-07-5287 | cs.DS | http://creativecommons.org/licenses/publicdomain/ | There exists an injective, information-preserving function that maps a
semantic network (i.e a directed labeled network) to a directed network (i.e. a
directed unlabeled network). The edge label in the semantic network is
represented as a topological feature of the directed network. Also, there
exists an injective function that maps a directed network to an undirected
network (i.e. an undirected unlabeled network). The edge directionality in the
directed network is represented as a topological feature of the undirected
network. Through function composition, there exists an injective function that
maps a semantic network to an undirected network. Thus, aside from space
constraints, the semantic network construct does not have any modeling
functionality that is not possible with either a directed or undirected network
representation. Two proofs of this idea will be presented. The first is a proof
of the aforementioned function composition concept. The second is a simpler
proof involving an undirected binary encoding of a semantic network.
| [
{
"version": "v1",
"created": "Wed, 2 Apr 2008 01:19:55 GMT"
}
] | 2008-12-02T00:00:00 | [
[
"Rodriguez",
"Marko A.",
""
]
] |
0804.0362 | Lenka Zdeborova | John Ardelius, Lenka Zdeborov\'a | Exhaustive enumeration unveils clustering and freezing in random 3-SAT | 4 pages, 3 figures | Phys. Rev. E 78, 040101(R) (2008) | 10.1103/PhysRevE.78.040101 | null | cond-mat.stat-mech cond-mat.dis-nn cs.CC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study geometrical properties of the complete set of solutions of the
random 3-satisfiability problem. We show that even for moderate system sizes
the number of clusters corresponds surprisingly well with the theoretic
asymptotic prediction. We locate the freezing transition in the space of
solutions which has been conjectured to be relevant in explaining the onset of
computational hardness in random constraint satisfaction problems.
| [
{
"version": "v1",
"created": "Wed, 2 Apr 2008 14:32:44 GMT"
},
{
"version": "v2",
"created": "Mon, 14 Apr 2008 09:31:46 GMT"
}
] | 2008-10-02T00:00:00 | [
[
"Ardelius",
"John",
""
],
[
"Zdeborová",
"Lenka",
""
]
] |
0804.0570 | Daniel Raible | Jianer Chen, Henning Fernau, Dan Ning, Daniel Raible, Jianxin Wang | A Parameterized Perspective on $P_2$-Packings | null | null | null | null | cs.DS cs.CC cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | }We study (vertex-disjoint) $P_2$-packings in graphs under a parameterized
perspective. Starting from a maximal $P_2$-packing $\p$ of size $j$ we use
extremal arguments for determining how many vertices of $\p$ appear in some
$P_2$-packing of size $(j+1)$. We basically can 'reuse' $2.5j$ vertices. We
also present a kernelization algorithm that gives a kernel of size bounded by
$7k$. With these two results we build an algorithm which constructs a
$P_2$-packing of size $k$ in time $\Oh^*(2.482^{3k})$.
| [
{
"version": "v1",
"created": "Thu, 3 Apr 2008 14:36:19 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Chen",
"Jianer",
""
],
[
"Fernau",
"Henning",
""
],
[
"Ning",
"Dan",
""
],
[
"Raible",
"Daniel",
""
],
[
"Wang",
"Jianxin",
""
]
] |
0804.0577 | Oskar Sandberg | Oskar Sandberg | Decentralized Search with Random Costs | null | null | null | null | math.PR cs.DS | http://creativecommons.org/licenses/by-nc-sa/3.0/ | A decentralized search algorithm is a method of routing on a random graph
that uses only limited, local, information about the realization of the graph.
In some random graph models it is possible to define such algorithms which
produce short paths when routing from any vertex to any other, while for others
it is not.
We consider random graphs with random costs assigned to the edges. In this
situation, we use the methods of stochastic dynamic programming to create a
decentralized search method which attempts to minimize the total cost, rather
than the number of steps, of each path. We show that it succeeds in doing so
among all decentralized search algorithms which monotonically approach the
destination. Our algorithm depends on knowing the expected cost of routing from
every vertex to any other, but we show that this may be calculated iteratively,
and in practice can be easily estimated from the cost of previous routes and
compressed into a small routing table. The methods applied here can also be
applied directly in other situations, such as efficient searching in graphs
with varying vertex degrees.
| [
{
"version": "v1",
"created": "Thu, 3 Apr 2008 15:32:29 GMT"
}
] | 2008-04-04T00:00:00 | [
[
"Sandberg",
"Oskar",
""
]
] |
0804.0722 | Daniel Karapetyan | Gregory Gutin, Daniel Karapetyan | A Memetic Algorithm for the Generalized Traveling Salesman Problem | 15 pages, to appear in Natural Computing, Springer, available online:
http://www.springerlink.com/content/5v4568l492272865/?p=e1779dd02e4d4cbfa49d0d27b19b929f&pi=13 | Natural Computing 9(1) (2010) 47-60 | 10.1007/s11047-009-9111-6 | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The generalized traveling salesman problem (GTSP) is an extension of the
well-known traveling salesman problem. In GTSP, we are given a partition of
cities into groups and we are required to find a minimum length tour that
includes exactly one city from each group. The recent studies on this subject
consider different variations of a memetic algorithm approach to the GTSP. The
aim of this paper is to present a new memetic algorithm for GTSP with a
powerful local search procedure. The experiments show that the proposed
algorithm clearly outperforms all of the known heuristics with respect to both
solution quality and running time. While the other memetic algorithms were
designed only for the symmetric GTSP, our algorithm can solve both symmetric
and asymmetric instances.
| [
{
"version": "v1",
"created": "Fri, 4 Apr 2008 13:21:40 GMT"
},
{
"version": "v2",
"created": "Tue, 11 Nov 2008 23:58:20 GMT"
},
{
"version": "v3",
"created": "Fri, 13 Mar 2009 22:13:27 GMT"
}
] | 2010-03-30T00:00:00 | [
[
"Gutin",
"Gregory",
""
],
[
"Karapetyan",
"Daniel",
""
]
] |
0804.0735 | Daniel Karapetyan | Gregory Gutin and Daniel Karapetyan | Generalized Traveling Salesman Problem Reduction Algorithms | To appear in Algorithmic Operations Research | Algorithmic Operations Research 4 (2009) 144-154 | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The generalized traveling salesman problem (GTSP) is an extension of the
well-known traveling salesman problem. In GTSP, we are given a partition of
cities into groups and we are required to find a minimum length tour that
includes exactly one city from each group. The aim of this paper is to present
a problem reduction algorithm that deletes redundant vertices and edges,
preserving the optimal solution. The algorithm's running time is O(N^3) in the
worst case, but it is significantly faster in practice. The algorithm has
reduced the problem size by 15-20% on average in our experiments and this has
decreased the solution time by 10-60% for each of the considered solvers.
| [
{
"version": "v1",
"created": "Fri, 4 Apr 2008 13:36:19 GMT"
},
{
"version": "v2",
"created": "Tue, 14 Apr 2009 17:36:47 GMT"
}
] | 2010-03-30T00:00:00 | [
[
"Gutin",
"Gregory",
""
],
[
"Karapetyan",
"Daniel",
""
]
] |
0804.0743 | Fabien Mathieu | Laurent Viennot (INRIA Rocquencourt), Yacine Boufkhad (INRIA
Rocquencourt, LIAFA), Fabien Mathieu (INRIA Rocquencourt, FT R&D), Fabien De
Montgolfier (INRIA Rocquencourt, LIAFA), Diego Perino (INRIA Rocquencourt, FT
R&D) | Scalable Distributed Video-on-Demand: Theoretical Bounds and Practical
Algorithms | null | null | null | RR-6496 | cs.NI cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We analyze a distributed system where n nodes called boxes store a large set
of videos and collaborate to serve simultaneously n videos or less. We explore
under which conditions such a system can be scalable while serving any sequence
of demands. We model this problem through a combination of two algorithms: a
video allocation algorithm and a connection scheduling algorithm. The latter
plays against an adversary that incrementally proposes video requests.
| [
{
"version": "v1",
"created": "Fri, 4 Apr 2008 14:08:49 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Apr 2008 07:16:36 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Viennot",
"Laurent",
"",
"INRIA Rocquencourt"
],
[
"Boufkhad",
"Yacine",
"",
"INRIA\n Rocquencourt, LIAFA"
],
[
"Mathieu",
"Fabien",
"",
"INRIA Rocquencourt, FT R&D"
],
[
"De Montgolfier",
"Fabien",
"",
"INRIA Rocquencourt, LIAFA"
],
[
"Perino",
"Diego",
"",
"INRIA Rocquencourt, FT\n R&D"
]
] |
0804.0936 | Shripad Thite | Mark de Berg and Shripad Thite | Cache-Oblivious Selection in Sorted X+Y Matrices | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Let X[0..n-1] and Y[0..m-1] be two sorted arrays, and define the mxn matrix A
by A[j][i]=X[i]+Y[j]. Frederickson and Johnson gave an efficient algorithm for
selecting the k-th smallest element from A. We show how to make this algorithm
IO-efficient. Our cache-oblivious algorithm performs O((m+n)/B) IOs, where B is
the block size of memory transfers.
| [
{
"version": "v1",
"created": "Sun, 6 Apr 2008 22:31:04 GMT"
}
] | 2008-04-08T00:00:00 | [
[
"de Berg",
"Mark",
""
],
[
"Thite",
"Shripad",
""
]
] |
0804.0940 | Shripad Thite | Shripad Thite | Optimum Binary Search Trees on the Hierarchical Memory Model | M.S. thesis; Department of Computer Science, University of Illinois
at Urbana-Champaign; CSL Technical Report UILU-ENG-00-2215 ACT-142; November
2000 | null | null | UILU-ENG-00-2215 ACT-142 | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Hierarchical Memory Model (HMM) of computation is similar to the standard
Random Access Machine (RAM) model except that the HMM has a non-uniform memory
organized in a hierarchy of levels numbered 1 through h. The cost of accessing
a memory location increases with the level number, and accesses to memory
locations belonging to the same level cost the same. Formally, the cost of a
single access to the memory location at address a is given by m(a), where m: N
-> N is the memory cost function, and the h distinct values of m model the
different levels of the memory hierarchy.
We study the problem of constructing and storing a binary search tree (BST)
of minimum cost, over a set of keys, with probabilities for successful and
unsuccessful searches, on the HMM with an arbitrary number of memory levels,
and for the special case h=2.
While the problem of constructing optimum binary search trees has been well
studied for the standard RAM model, the additional parameter m for the HMM
increases the combinatorial complexity of the problem. We present two dynamic
programming algorithms to construct optimum BSTs bottom-up. These algorithms
run efficiently under some natural assumptions about the memory hierarchy. We
also give an efficient algorithm to construct a BST that is close to optimum,
by modifying a well-known linear-time approximation algorithm for the RAM
model. We conjecture that the problem of constructing an optimum BST for the
HMM with an arbitrary memory cost function m is NP-complete.
| [
{
"version": "v1",
"created": "Mon, 7 Apr 2008 00:06:08 GMT"
}
] | 2008-04-08T00:00:00 | [
[
"Thite",
"Shripad",
""
]
] |
0804.1115 | Oskar Sandberg | Olof Mogren, Oskar Sandberg, Vilhelm Verendel and Devdatt Dubhashi | Adaptive Dynamics of Realistic Small-World Networks | null | null | null | null | cs.DS cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Continuing in the steps of Jon Kleinberg's and others celebrated work on
decentralized search in small-world networks, we conduct an experimental
analysis of a dynamic algorithm that produces small-world networks. We find
that the algorithm adapts robustly to a wide variety of situations in realistic
geographic networks with synthetic test data and with real world data, even
when vertices are uneven and non-homogeneously distributed.
We investigate the same algorithm in the case where some vertices are more
popular destinations for searches than others, for example obeying power-laws.
We find that the algorithm adapts and adjusts the networks according to the
distributions, leading to improved performance. The ability of the dynamic
process to adapt and create small worlds in such diverse settings suggests a
possible mechanism by which such networks appear in nature.
| [
{
"version": "v1",
"created": "Mon, 7 Apr 2008 19:39:59 GMT"
}
] | 2008-04-08T00:00:00 | [
[
"Mogren",
"Olof",
""
],
[
"Sandberg",
"Oskar",
""
],
[
"Verendel",
"Vilhelm",
""
],
[
"Dubhashi",
"Devdatt",
""
]
] |
0804.1170 | Daniel \v{S}tefankovi\v{c} | Satyaki Mahalanabis, Daniel Stefankovic | Approximating L1-distances between mixture distributions using random
projections | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider the problem of computing L1-distances between every pair
ofcprobability densities from a given family. We point out that the technique
of Cauchy random projections (Indyk'06) in this context turns into stochastic
integrals with respect to Cauchy motion.
For piecewise-linear densities these integrals can be sampled from if one can
sample from the stochastic integral of the function x->(1,x). We give an
explicit density function for this stochastic integral and present an efficient
sampling algorithm. As a consequence we obtain an efficient algorithm to
approximate the L1-distances with a small relative error.
For piecewise-polynomial densities we show how to approximately sample from
the distributions resulting from the stochastic integrals. This also results in
an efficient algorithm to approximate the L1-distances, although our inability
to get exact samples worsens the dependence on the parameters.
| [
{
"version": "v1",
"created": "Tue, 8 Apr 2008 02:11:13 GMT"
}
] | 2008-04-09T00:00:00 | [
[
"Mahalanabis",
"Satyaki",
""
],
[
"Stefankovic",
"Daniel",
""
]
] |
0804.1409 | Murat Ali Bayir Mr. | Murat Ali Bayir, Ismail Hakki Toroslu, Ahmet Cosar, Guven Fidan | Discovering More Accurate Frequent Web Usage Patterns | 19 pages, 6 figures | null | null | null | cs.DB cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Web usage mining is a type of web mining, which exploits data mining
techniques to discover valuable information from navigation behavior of World
Wide Web users. As in classical data mining, data preparation and pattern
discovery are the main issues in web usage mining. The first phase of web usage
mining is the data processing phase, which includes the session reconstruction
operation from server logs. Session reconstruction success directly affects the
quality of the frequent patterns discovered in the next phase. In reactive web
usage mining techniques, the source data is web server logs and the topology of
the web pages served by the web server domain. Other kinds of information
collected during the interactive browsing of web site by user, such as cookies
or web logs containing similar information, are not used. The next phase of web
usage mining is discovering frequent user navigation patterns. In this phase,
pattern discovery methods are applied on the reconstructed sessions obtained in
the first phase in order to discover frequent user patterns. In this paper, we
propose a frequent web usage pattern discovery method that can be applied after
session reconstruction phase. In order to compare accuracy performance of
session reconstruction phase and pattern discovery phase, we have used an agent
simulator, which models behavior of web users and generates web user navigation
as well as the log data kept by the web server.
| [
{
"version": "v1",
"created": "Wed, 9 Apr 2008 05:46:26 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Bayir",
"Murat Ali",
""
],
[
"Toroslu",
"Ismail Hakki",
""
],
[
"Cosar",
"Ahmet",
""
],
[
"Fidan",
"Guven",
""
]
] |
0804.1724 | Sudipto Guha | Sudipto Guha and Kamesh Munagala and Saswati Sarkar | Information Acquisition and Exploitation in Multichannel Wireless
Networks | 29 pages | null | null | null | cs.DS cs.NI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A wireless system with multiple channels is considered, where each channel
has several transmission states. A user learns about the instantaneous state of
an available channel by transmitting a control packet in it. Since probing all
channels consumes significant energy and time, a user needs to determine what
and how much information it needs to acquire about the instantaneous states of
the available channels so that it can maximize its transmission rate. This
motivates the study of the trade-off between the cost of information
acquisition and its value towards improving the transmission rate.
A simple model is presented for studying this information acquisition and
exploitation trade-off when the channels are multi-state, with different
distributions and information acquisition costs. The objective is to maximize a
utility function which depends on both the cost and value of information.
Solution techniques are presented for computing near-optimal policies with
succinct representation in polynomial time. These policies provably achieve at
least a fixed constant factor of the optimal utility on any problem instance,
and in addition, have natural characterizations. The techniques are based on
exploiting the structure of the optimal policy, and use of Lagrangean
relaxations which simplify the space of approximately optimal solutions.
| [
{
"version": "v1",
"created": "Thu, 10 Apr 2008 14:53:30 GMT"
}
] | 2008-04-11T00:00:00 | [
[
"Guha",
"Sudipto",
""
],
[
"Munagala",
"Kamesh",
""
],
[
"Sarkar",
"Saswati",
""
]
] |
0804.1845 | Ely Porat | Ely Porat | An Optimal Bloom Filter Replacement Based on Matrix Solving | A lectureon this paper will be available in Google video | null | null | null | cs.DS cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We suggest a method for holding a dictionary data structure, which maps keys
to values, in the spirit of Bloom Filters. The space requirements of the
dictionary we suggest are much smaller than those of a hashtable. We allow
storing n keys, each mapped to value which is a string of k bits. Our suggested
method requires nk + o(n) bits space to store the dictionary, and O(n) time to
produce the data structure, and allows answering a membership query in O(1)
memory probes. The dictionary size does not depend on the size of the keys.
However, reducing the space requirements of the data structure comes at a
certain cost. Our dictionary has a small probability of a one sided error. When
attempting to obtain the value for a key that is stored in the dictionary we
always get the correct answer. However, when testing for membership of an
element that is not stored in the dictionary, we may get an incorrect answer,
and when requesting the value of such an element we may get a certain random
value. Our method is based on solving equations in GF(2^k) and using several
hash functions. Another significant advantage of our suggested method is that
we do not require using sophisticated hash functions. We only require pairwise
independent hash functions. We also suggest a data structure that requires only
nk bits space, has O(n2) preprocessing time, and has a O(log n) query time.
However, this data structures requires a uniform hash functions. In order
replace a Bloom Filter of n elements with an error proability of 2^{-k}, we
require nk + o(n) memory bits, O(1) query time, O(n) preprocessing time, and
only pairwise independent hash function. Even the most advanced previously
known Bloom Filter would require nk+O(n) space, and a uniform hash functions,
so our method is significantly less space consuming especially when k is small.
| [
{
"version": "v1",
"created": "Fri, 11 Apr 2008 11:24:04 GMT"
}
] | 2008-04-14T00:00:00 | [
[
"Porat",
"Ely",
""
]
] |
0804.1888 | Latorre | Frank Verstraete, J. Ignacio Cirac, Jose I. Latorre | Quantum circuits for strongly correlated quantum systems | null | null | 10.1103/PhysRevA.79.032316 | null | quant-ph cond-mat.str-el cs.DS hep-th | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In recent years, we have witnessed an explosion of experimental tools by
which quantum systems can be manipulated in a controlled and coherent way. One
of the most important goals now is to build quantum simulators, which would
open up the possibility of exciting experiments probing various theories in
regimes that are not achievable under normal lab circumstances. Here we present
a novel approach to gain detailed control on the quantum simulation of strongly
correlated quantum many-body systems by constructing the explicit quantum
circuits that diagonalize their dynamics. We show that the exact quantum
circuits underlying some of the most relevant many-body Hamiltonians only need
a finite amount of local gates. As a particularly simple instance, the full
dynamics of a one-dimensional Quantum Ising model in a transverse field with
four spins is shown to be reproduced using a quantum circuit of only six local
gates. This opens up the possibility of experimentally producing strongly
correlated states, their time evolution at zero time and even thermal
superpositions at zero temperature. Our method also allows to uncover the exact
circuits corresponding to models that exhibit topological order and to
stabilizer states.
| [
{
"version": "v1",
"created": "Fri, 11 Apr 2008 12:52:44 GMT"
}
] | 2009-11-13T00:00:00 | [
[
"Verstraete",
"Frank",
""
],
[
"Cirac",
"J. Ignacio",
""
],
[
"Latorre",
"Jose I.",
""
]
] |
0804.2032 | Frederic Dorn Harald | Paul Bonsma and Frederic Dorn | Tight Bounds and Faster Algorithms for Directed Max-Leaf Problems | 17 pages, 6 figures | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | An out-tree $T$ of a directed graph $D$ is a rooted tree subgraph with all
arcs directed outwards from the root. An out-branching is a spanning out-tree.
By $l(D)$ and $l_s(D)$ we denote the maximum number of leaves over all
out-trees and out-branchings of $D$, respectively.
We give fixed parameter tractable algorithms for deciding whether $l_s(D)\geq
k$ and whether $l(D)\geq k$ for a digraph $D$ on $n$ vertices, both with time
complexity $2^{O(k\log k)} \cdot n^{O(1)}$. This improves on previous
algorithms with complexity $2^{O(k^3\log k)} \cdot n^{O(1)}$ and $2^{O(k\log^2
k)} \cdot n^{O(1)}$, respectively.
To obtain the complexity bound in the case of out-branchings, we prove that
when all arcs of $D$ are part of at least one out-branching, $l_s(D)\geq
l(D)/3$. The second bound we prove in this paper states that for strongly
connected digraphs $D$ with minimum in-degree 3, $l_s(D)\geq \Theta(\sqrt{n})$,
where previously $l_s(D)\geq \Theta(\sqrt[3]{n})$ was the best known bound.
This bound is tight, and also holds for the larger class of digraphs with
minimum in-degree 3 in which every arc is part of at least one out-branching.
| [
{
"version": "v1",
"created": "Sat, 12 Apr 2008 20:50:59 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Bonsma",
"Paul",
""
],
[
"Dorn",
"Frederic",
""
]
] |
0804.2097 | Tim Roughgarden | Jason D. Hartline and Tim Roughgarden | Optimal Mechansim Design and Money Burning | 23 pages, 1 figure | null | null | null | cs.GT cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Mechanism design is now a standard tool in computer science for aligning the
incentives of self-interested agents with the objectives of a system designer.
There is, however, a fundamental disconnect between the traditional application
domains of mechanism design (such as auctions) and those arising in computer
science (such as networks): while monetary transfers (i.e., payments) are
essential for most of the known positive results in mechanism design, they are
undesirable or even technologically infeasible in many computer systems.
Classical impossibility results imply that the reach of mechanisms without
transfers is severely limited.
Computer systems typically do have the ability to reduce service
quality--routing systems can drop or delay traffic, scheduling protocols can
delay the release of jobs, and computational payment schemes can require
computational payments from users (e.g., in spam-fighting systems). Service
degradation is tantamount to requiring that users burn money}, and such
``payments'' can be used to influence the preferences of the agents at a cost
of degrading the social surplus.
We develop a framework for the design and analysis of money-burning
mechanisms to maximize the residual surplus--the total value of the chosen
outcome minus the payments required.
| [
{
"version": "v1",
"created": "Mon, 14 Apr 2008 04:32:45 GMT"
}
] | 2008-04-15T00:00:00 | [
[
"Hartline",
"Jason D.",
""
],
[
"Roughgarden",
"Tim",
""
]
] |
0804.2112 | Shai Gutner | Yossi Azar, Iftah Gamzu and Shai Gutner | Truthful Unsplittable Flow for Large Capacity Networks | null | Proc. of 19th SPAA (2007), 320-329 | null | null | cs.DS cs.GT | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we focus our attention on the large capacities unsplittable
flow problem in a game theoretic setting. In this setting, there are selfish
agents, which control some of the requests characteristics, and may be
dishonest about them. It is worth noting that in game theoretic settings many
standard techniques, such as randomized rounding, violate certain monotonicity
properties, which are imperative for truthfulness, and therefore cannot be
employed. In light of this state of affairs, we design a monotone deterministic
algorithm, which is based on a primal-dual machinery, which attains an
approximation ratio of $\frac{e}{e-1}$, up to a disparity of $\epsilon$ away.
This implies an improvement on the current best truthful mechanism, as well as
an improvement on the current best combinatorial algorithm for the problem
under consideration. Surprisingly, we demonstrate that any algorithm in the
family of reasonable iterative path minimizing algorithms, cannot yield a
better approximation ratio. Consequently, it follows that in order to achieve a
monotone PTAS, if exists, one would have to exert different techniques. We also
consider the large capacities \textit{single-minded multi-unit combinatorial
auction problem}. This problem is closely related to the unsplittable flow
problem since one can formulate it as a special case of the integer linear
program of the unsplittable flow problem. Accordingly, we obtain a comparable
performance guarantee by refining the algorithm suggested for the unsplittable
flow problem.
| [
{
"version": "v1",
"created": "Mon, 14 Apr 2008 08:03:30 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Azar",
"Yossi",
""
],
[
"Gamzu",
"Iftah",
""
],
[
"Gutner",
"Shai",
""
]
] |
0804.2288 | Shipra Agrawal | Shipra Agrawal, Zizhuo Wang, Yinyu Ye | Parimutuel Betting on Permutations | null | null | null | null | cs.GT cs.CC cs.DS cs.MA | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We focus on a permutation betting market under parimutuel call auction model
where traders bet on the final ranking of n candidates. We present a
Proportional Betting mechanism for this market. Our mechanism allows the
traders to bet on any subset of the n x n 'candidate-rank' pairs, and rewards
them proportionally to the number of pairs that appear in the final outcome. We
show that market organizer's decision problem for this mechanism can be
formulated as a convex program of polynomial size. More importantly, the
formulation yields a set of n x n unique marginal prices that are sufficient to
price the bets in this mechanism, and are computable in polynomial-time. The
marginal prices reflect the traders' beliefs about the marginal distributions
over outcomes. We also propose techniques to compute the joint distribution
over n! permutations from these marginal distributions. We show that using a
maximum entropy criterion, we can obtain a concise parametric form (with only n
x n parameters) for the joint distribution which is defined over an
exponentially large state space. We then present an approximation algorithm for
computing the parameters of this distribution. In fact, the algorithm addresses
the generic problem of finding the maximum entropy distribution over
permutations that has a given mean, and may be of independent interest.
| [
{
"version": "v1",
"created": "Tue, 15 Apr 2008 00:20:17 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Agrawal",
"Shipra",
""
],
[
"Wang",
"Zizhuo",
""
],
[
"Ye",
"Yinyu",
""
]
] |
0804.2699 | Dennis Huo | Ian Christopher, Dennis Huo, and Bryan Jacobs | A Critique of a Polynomial-time SAT Solver Devised by Sergey Gubin | null | null | null | null | cs.CC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper refutes the validity of the polynomial-time algorithm for solving
satisfiability proposed by Sergey Gubin. Gubin introduces the algorithm using
3-SAT and eventually expands it to accept a broad range of forms of the Boolean
satisfiability problem. Because 3-SAT is NP-complete, the algorithm would have
implied P = NP, had it been correct. Additionally, this paper refutes the
correctness of his polynomial-time reduction of SAT to 2-SAT.
| [
{
"version": "v1",
"created": "Wed, 16 Apr 2008 23:00:51 GMT"
}
] | 2008-04-18T00:00:00 | [
[
"Christopher",
"Ian",
""
],
[
"Huo",
"Dennis",
""
],
[
"Jacobs",
"Bryan",
""
]
] |
0804.3028 | Saket Saurabh | Michael Fellows, Fedor Fomin, Daniel Lokshtanov, Elena Losievskaja,
Frances A. Rosamond and Saket Saurabh | Parameterized Low-distortion Embeddings - Graph metrics into lines and
trees | 19 pages, 1 Figure | null | null | null | cs.DS cs.CC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We revisit the issue of low-distortion embedding of metric spaces into the
line, and more generally, into the shortest path metric of trees, from the
parameterized complexity perspective.Let $M=M(G)$ be the shortest path metric
of an edge weighted graph $G=(V,E)$ on $n$ vertices. We describe algorithms for
the problem of finding a low distortion non-contracting embedding of $M$ into
line and tree metrics.
We give an $O(nd^4(2d+1)^{2d})$ time algorithm that for an unweighted graph
metric $M$ and integer $d$ either constructs an embedding of $M$ into the line
with distortion at most $d$, or concludes that no such embedding exists. We
find the result surprising, because the considered problem bears a strong
resemblance to the notoriously hard Bandwidth Minimization problem which does
not admit any FPT algorithm unless an unlikely collapse of parameterized
complexity classes occurs.
We show that our algorithm can also be applied to construct small distortion
embeddings of weighted graph metrics. The running time of our algorithm is
$O(n(dW)^4(2d+1)^{2dW})$ where $W$ is the largest edge weight of the input
graph. We also show that deciding whether a weighted graph metric $M(G)$ with
maximum weight $W < |V(G)|$ can be embedded into the line with distortion at
most $d$ is NP-Complete for every fixed rational $d \geq 2$. This rules out any
possibility of an algorithm with running time $O((nW)^{h(d)})$ where $h$ is a
function of $d$ alone.
We generalize the result on embedding into the line by proving that for any
tree $T$ with maximum degree $\Delta$, embedding of $M$ into a shortest path
metric of $T$ is FPT, parameterized by $(\Delta,d)$.
| [
{
"version": "v1",
"created": "Fri, 18 Apr 2008 14:39:41 GMT"
}
] | 2008-04-21T00:00:00 | [
[
"Fellows",
"Michael",
""
],
[
"Fomin",
"Fedor",
""
],
[
"Lokshtanov",
"Daniel",
""
],
[
"Losievskaja",
"Elena",
""
],
[
"Rosamond",
"Frances A.",
""
],
[
"Saurabh",
"Saket",
""
]
] |
0804.3615 | Jarek Duda | Jarek Duda | Combinatorial invariants for graph isomorphism problem | null | null | null | null | cs.CC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Presented approach in polynomial time calculates large number of invariants
for each vertex, which won't change with graph isomorphism and should fully
determine the graph. For example numbers of closed paths of length k for given
starting vertex, what can be though as the diagonal terms of k-th power of the
adjacency matrix. For k=2 we would get degree of verities invariant, higher
describes local topology deeper. Now if two graphs are isomorphic, they have
the same set of such vectors of invariants - we can sort theses vectors
lexicographically and compare them. If they agree, permutations from sorting
allow to reconstruct the isomorphism. I'm presenting arguments that these
invariants should fully determine the graph, but unfortunately I can't prove it
in this moment. This approach can give hope, that maybe P=NP - instead of
checking all instances, we should make arithmetics on these large numbers.
| [
{
"version": "v1",
"created": "Tue, 22 Apr 2008 22:16:46 GMT"
},
{
"version": "v2",
"created": "Wed, 23 Apr 2008 21:54:08 GMT"
},
{
"version": "v3",
"created": "Fri, 9 May 2008 07:09:06 GMT"
},
{
"version": "v4",
"created": "Mon, 19 May 2008 14:20:36 GMT"
}
] | 2008-05-19T00:00:00 | [
[
"Duda",
"Jarek",
""
]
] |
0804.3860 | Hsiao-Fei Liu | Hsiao-Fei Liu and Kun-Mao Chao | An $\tilde{O}(n^{2.5})$-Time Algorithm for Online Topological Ordering | Better results have been proposed in the following paper: Haeupler,
Kavitha, Mathew, Sen, Tarjan: Faster Algorithms for Incremental Topological
Ordering. ICALP (1) 2008: 421-433 | null | null | null | cs.GT cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present an $\tilde{O}(n^{2.5})$-time algorithm for maintaining the
topological order of a directed acyclic graph with $n$ vertices while inserting
$m$ edges.
| [
{
"version": "v1",
"created": "Thu, 24 Apr 2008 09:40:20 GMT"
},
{
"version": "v2",
"created": "Sat, 23 Aug 2008 07:10:20 GMT"
}
] | 2008-08-23T00:00:00 | [
[
"Liu",
"Hsiao-Fei",
""
],
[
"Chao",
"Kun-Mao",
""
]
] |
0804.3902 | Tiziana Calamoneri | Tiziana Calamoneri, Andrea E.F. Clementi, Angelo Monti, Gianluca
Rossi, Riccardo Silvestri | Minimum-energy broadcast in random-grid ad-hoc networks: approximation
and distributed algorithms | 13 pages, 3 figures, 1 table | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Min Energy broadcast problem consists in assigning transmission ranges to
the nodes of an ad-hoc network in order to guarantee a directed spanning tree
from a given source node and, at the same time, to minimize the energy
consumption (i.e. the energy cost) yielded by the range assignment. Min energy
broadcast is known to be NP-hard.
We consider random-grid networks where nodes are chosen independently at
random from the $n$ points of a $\sqrt n \times \sqrt n$ square grid in the
plane. The probability of the existence of a node at a given point of the grid
does depend on that point, that is, the probability distribution can be
non-uniform.
By using information-theoretic arguments, we prove a lower bound
$(1-\epsilon) \frac n{\pi}$ on the energy cost of any feasible solution for
this problem. Then, we provide an efficient solution of energy cost not larger
than $1.1204 \frac n{\pi}$.
Finally, we present a fully-distributed protocol that constructs a broadcast
range assignment of energy cost not larger than $8n$,thus still yielding
constant approximation. The energy load is well balanced and, at the same time,
the work complexity (i.e. the energy due to all message transmissions of the
protocol) is asymptotically optimal. The completion time of the protocol is
only an $O(\log n)$ factor slower than the optimum. The approximation quality
of our distributed solution is also experimentally evaluated.
All bounds hold with probability at least $1-1/n^{\Theta(1)}$.
| [
{
"version": "v1",
"created": "Thu, 24 Apr 2008 11:17:57 GMT"
}
] | 2008-04-25T00:00:00 | [
[
"Calamoneri",
"Tiziana",
""
],
[
"Clementi",
"Andrea E. F.",
""
],
[
"Monti",
"Angelo",
""
],
[
"Rossi",
"Gianluca",
""
],
[
"Silvestri",
"Riccardo",
""
]
] |
0804.3947 | Peter Sanders | Peter Sanders | Time Dependent Contraction Hierarchies -- Basic Algorithmic Ideas | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Contraction hierarchies are a simple hierarchical routing technique that has
proved extremely efficient for static road networks. We explain how to
generalize them to networks with time-dependent edge weights. This is the first
hierarchical speedup technique for time-dependent routing that allows
bidirectional query algorithms.
| [
{
"version": "v1",
"created": "Thu, 24 Apr 2008 15:24:08 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Sanders",
"Peter",
""
]
] |
0804.4039 | Ioannis Chatzigiannakis | Ioannis Chatzigiannakis, Georgios Giannoulis and Paul G. Spirakis | Energy and Time Efficient Scheduling of Tasks with Dependencies on
Asymmetric Multiprocessors | null | null | null | RACTI-RU1-2008-10 | cs.DC cs.DS cs.PF | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work we study the problem of scheduling tasks with dependencies in
multiprocessor architectures where processors have different speeds. We present
the preemptive algorithm "Save-Energy" that given a schedule of tasks it post
processes it to improve the energy efficiency without any deterioration of the
makespan. In terms of time efficiency, we show that preemptive scheduling in an
asymmetric system can achieve the same or better optimal makespan than in a
symmetric system. Motivited by real multiprocessor systems, we investigate
architectures that exhibit limited asymmetry: there are two essentially
different speeds. Interestingly, this special case has not been studied in the
field of parallel computing and scheduling theory; only the general case was
studied where processors have $K$ essentially different speeds. We present the
non-preemptive algorithm ``Remnants'' that achieves almost optimal makespan. We
provide a refined analysis of a recent scheduling method. Based on this
analysis, we specialize the scheduling policy and provide an algorithm of $(3 +
o(1))$ expected approximation factor. Note that this improves the previous best
factor (6 for two speeds). We believe that our work will convince researchers
to revisit this well studied scheduling problem for these simple, yet
realistic, asymmetric multiprocessor architectures.
| [
{
"version": "v1",
"created": "Fri, 25 Apr 2008 03:16:21 GMT"
},
{
"version": "v2",
"created": "Fri, 6 Jun 2008 14:21:18 GMT"
}
] | 2008-06-09T00:00:00 | [
[
"Chatzigiannakis",
"Ioannis",
""
],
[
"Giannoulis",
"Georgios",
""
],
[
"Spirakis",
"Paul G.",
""
]
] |
0804.4138 | Jelani Nelson | Nicholas J. A. Harvey, Jelani Nelson, Krzysztof Onak | Sketching and Streaming Entropy via Approximation Theory | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We conclude a sequence of work by giving near-optimal sketching and streaming
algorithms for estimating Shannon entropy in the most general streaming model,
with arbitrary insertions and deletions. This improves on prior results that
obtain suboptimal space bounds in the general model, and near-optimal bounds in
the insertion-only model without sketching. Our high-level approach is simple:
we give algorithms to estimate Renyi and Tsallis entropy, and use them to
extrapolate an estimate of Shannon entropy. The accuracy of our estimates is
proven using approximation theory arguments and extremal properties of
Chebyshev polynomials, a technique which may be useful for other problems. Our
work also yields the best-known and near-optimal additive approximations for
entropy, and hence also for conditional entropy and mutual information.
| [
{
"version": "v1",
"created": "Fri, 25 Apr 2008 16:04:20 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Harvey",
"Nicholas J. A.",
""
],
[
"Nelson",
"Jelani",
""
],
[
"Onak",
"Krzysztof",
""
]
] |
0804.4666 | Anna Gilbert | R. Berinde, A. C. Gilbert, P. Indyk, H. Karloff, M. J. Strauss | Combining geometry and combinatorics: A unified approach to sparse
signal recovery | null | null | null | null | cs.DM cs.DS cs.NA | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | There are two main algorithmic approaches to sparse signal recovery:
geometric and combinatorial. The geometric approach starts with a geometric
constraint on the measurement matrix and then uses linear programming to decode
information about the signal from its measurements. The combinatorial approach
constructs the measurement matrix and a combinatorial decoding algorithm to
match. We present a unified approach to these two classes of sparse signal
recovery algorithms.
The unifying elements are the adjacency matrices of high-quality unbalanced
expanders. We generalize the notion of Restricted Isometry Property (RIP),
crucial to compressed sensing results for signal recovery, from the Euclidean
norm to the l_p norm for p about 1, and then show that unbalanced expanders are
essentially equivalent to RIP-p matrices.
From known deterministic constructions for such matrices, we obtain new
deterministic measurement matrix constructions and algorithms for signal
recovery which, compared to previous deterministic algorithms, are superior in
either the number of measurements or in noise tolerance.
| [
{
"version": "v1",
"created": "Tue, 29 Apr 2008 18:24:14 GMT"
}
] | 2008-04-30T00:00:00 | [
[
"Berinde",
"R.",
""
],
[
"Gilbert",
"A. C.",
""
],
[
"Indyk",
"P.",
""
],
[
"Karloff",
"H.",
""
],
[
"Strauss",
"M. J.",
""
]
] |
0804.4744 | V. Arvind | V. Arvind and Pushkar S. Joglekar | Lattice Problems, Gauge Functions and Parameterized Algorithms | null | null | null | null | cs.CC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Given a k-dimensional subspace M\subseteq \R^n and a full rank integer
lattice L\subseteq \R^n, the \emph{subspace avoiding problem} SAP is to find a
shortest vector in L\setminus M. Treating k as a parameter, we obtain new
parameterized approximation and exact algorithms for SAP based on the AKS
sieving technique. More precisely, we give a randomized
$(1+\epsilon)$-approximation algorithm for parameterized SAP that runs in time
2^{O(n)}.(1/\epsilon)^k, where the parameter k is the dimension of the subspace
M. Thus, we obtain a 2^{O(n)} time algorithm for \epsilon=2^{-O(n/k)}. We also
give a 2^{O(n+k\log k)} exact algorithm for the parameterized SAP for any
\ell_p norm.
Several of our algorithms work for all gauge functions as metric with some
natural restrictions, in particular for all \ell_p norms. We also prove an
\Omega(2^n) lower bound on the query complexity of AKS sieving based exact
algorithms for SVP that accesses the gauge function as oracle.
| [
{
"version": "v1",
"created": "Wed, 30 Apr 2008 06:39:21 GMT"
}
] | 2008-05-01T00:00:00 | [
[
"Arvind",
"V.",
""
],
[
"Joglekar",
"Pushkar S.",
""
]
] |
0804.4819 | Jukka Suomela | Michael A. Bender, S\'andor P. Fekete, Alexander Kr\"oller, Vincenzo
Liberatore, Joseph S. B. Mitchell, Valentin Polishchuk, Jukka Suomela | The Minimum Backlog Problem | 1+16 pages, 3 figures | Theoretical Computer Science 605 (2015), 51-61 | 10.1016/j.tcs.2015.08.027 | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study the minimum backlog problem (MBP). This online problem arises, e.g.,
in the context of sensor networks. We focus on two main variants of MBP.
The discrete MBP is a 2-person game played on a graph $G=(V,E)$. The player
is initially located at a vertex of the graph. In each time step, the adversary
pours a total of one unit of water into cups that are located on the vertices
of the graph, arbitrarily distributing the water among the cups. The player
then moves from her current vertex to an adjacent vertex and empties the cup at
that vertex. The player's objective is to minimize the backlog, i.e., the
maximum amount of water in any cup at any time.
The geometric MBP is a continuous-time version of the MBP: the cups are
points in the two-dimensional plane, the adversary pours water continuously at
a constant rate, and the player moves in the plane with unit speed. Again, the
player's objective is to minimize the backlog.
We show that the competitive ratio of any algorithm for the MBP has a lower
bound of $\Omega(D)$, where $D$ is the diameter of the graph (for the discrete
MBP) or the diameter of the point set (for the geometric MBP). Therefore we
focus on determining a strategy for the player that guarantees a uniform upper
bound on the absolute value of the backlog.
For the absolute value of the backlog there is a trivial lower bound of
$\Omega(D)$, and the deamortization analysis of Dietz and Sleator gives an
upper bound of $O(D\log N)$ for $N$ cups. Our main result is a tight upper
bound for the geometric MBP: we show that there is a strategy for the player
that guarantees a backlog of $O(D)$, independently of the number of cups.
| [
{
"version": "v1",
"created": "Wed, 30 Apr 2008 13:13:12 GMT"
},
{
"version": "v2",
"created": "Tue, 22 Mar 2016 20:54:15 GMT"
}
] | 2016-03-24T00:00:00 | [
[
"Bender",
"Michael A.",
""
],
[
"Fekete",
"Sándor P.",
""
],
[
"Kröller",
"Alexander",
""
],
[
"Liberatore",
"Vincenzo",
""
],
[
"Mitchell",
"Joseph S. B.",
""
],
[
"Polishchuk",
"Valentin",
""
],
[
"Suomela",
"Jukka",
""
]
] |
0804.4881 | Adolfo Piperno | Adolfo Piperno | Search Space Contraction in Canonical Labeling of Graphs | null | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The individualization-refinement paradigm for computing a canonical labeling
and the automorphism group of a graph is investigated. A new algorithmic design
aimed at reducing the size of the associated search space is introduced, and a
new tool, named "Traces", is presented, together with experimental results and
comparisons with existing software, such as McKay's "nauty". It is shown that
the approach presented here leads to a huge reduction in the search space,
thereby making computation feasible for several classes of graphs which are
hard for all the main canonical labeling tools in the literature.
| [
{
"version": "v1",
"created": "Wed, 30 Apr 2008 18:28:13 GMT"
},
{
"version": "v2",
"created": "Wed, 26 Jan 2011 15:52:11 GMT"
}
] | 2015-03-13T00:00:00 | [
[
"Piperno",
"Adolfo",
""
]
] |
0805.0389 | Chaitanya Swamy | Chaitanya Swamy | Algorithms for Probabilistically-Constrained Models of Risk-Averse
Stochastic Optimization with Black-Box Distributions | null | null | null | null | cs.DS cs.CC cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider various stochastic models that incorporate the notion of
risk-averseness into the standard 2-stage recourse model, and develop novel
techniques for solving the algorithmic problems arising in these models. A key
notable feature of our work that distinguishes it from work in some other
related models, such as the (standard) budget model and the (demand-) robust
model, is that we obtain results in the black-box setting, that is, where one
is given only sampling access to the underlying distribution. Our first model,
which we call the risk-averse budget model, incorporates the notion of
risk-averseness via a probabilistic constraint that restricts the probability
(according to the underlying distribution) with which the second-stage cost may
exceed a given budget B to at most a given input threshold \rho. We also a
consider a closely-related model that we call the risk-averse robust model,
where we seek to minimize the first-stage cost and the (1-\rho)-quantile of the
second-stage cost.
We obtain approximation algorithms for a variety of combinatorial
optimization problems including the set cover, vertex cover, multicut on trees,
min cut, and facility location problems, in the risk-averse budget and robust
models with black-box distributions. We obtain near-optimal solutions that
preserve the budget approximately and incur a small blow-up of the probability
threshold (both of which are unavoidable). To the best of our knowledge, these
are the first approximation results for problems involving probabilistic
constraints and black-box distributions. A major component of our results is a
fully polynomial approximation scheme for solving the LP-relaxation of the
risk-averse problem.
| [
{
"version": "v1",
"created": "Sun, 4 May 2008 03:57:52 GMT"
}
] | 2008-05-06T00:00:00 | [
[
"Swamy",
"Chaitanya",
""
]
] |
0805.0747 | Daniel Lemire | Hazel Webb, Owen Kaser, Daniel Lemire | Pruning Attribute Values From Data Cubes with Diamond Dicing | null | null | null | TR-08-011 (UNB Saint John) | cs.DB cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Data stored in a data warehouse are inherently multidimensional, but most
data-pruning techniques (such as iceberg and top-k queries) are unidimensional.
However, analysts need to issue multidimensional queries. For example, an
analyst may need to select not just the most profitable stores
or--separately--the most profitable products, but simultaneous sets of stores
and products fulfilling some profitability constraints. To fill this need, we
propose a new operator, the diamond dice. Because of the interaction between
dimensions, the computation of diamonds is challenging. We present the first
diamond-dicing experiments on large data sets. Experiments show that we can
compute diamond cubes over fact tables containing 100 million facts in less
than 35 minutes using a standard PC.
| [
{
"version": "v1",
"created": "Tue, 6 May 2008 15:45:15 GMT"
}
] | 2008-05-07T00:00:00 | [
[
"Webb",
"Hazel",
""
],
[
"Kaser",
"Owen",
""
],
[
"Lemire",
"Daniel",
""
]
] |
0805.0851 | Sebastien Tixeuil | Samuel Bernard (LIP6), St\'ephane Devismes (LRI), Maria Gradinariu
Potop-Butucaru (LIP6, INRIA Rocquencourt), S\'ebastien Tixeuil (LIP6) | Bounds for self-stabilization in unidirectional networks | null | null | null | RR-6524 | cs.DS cs.CC cs.DC cs.NI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A distributed algorithm is self-stabilizing if after faults and attacks hit
the system and place it in some arbitrary global state, the systems recovers
from this catastrophic situation without external intervention in finite time.
Unidirectional networks preclude many common techniques in self-stabilization
from being used, such as preserving local predicates. In this paper, we
investigate the intrinsic complexity of achieving self-stabilization in
unidirectional networks, and focus on the classical vertex coloring problem.
When deterministic solutions are considered, we prove a lower bound of $n$
states per process (where $n$ is the network size) and a recovery time of at
least $n(n-1)/2$ actions in total. We present a deterministic algorithm with
matching upper bounds that performs in arbitrary graphs. When probabilistic
solutions are considered, we observe that at least $\Delta + 1$ states per
process and a recovery time of $\Omega(n)$ actions in total are required (where
$\Delta$ denotes the maximal degree of the underlying simple undirected graph).
We present a probabilistically self-stabilizing algorithm that uses
$\mathtt{k}$ states per process, where $\mathtt{k}$ is a parameter of the
algorithm. When $\mathtt{k}=\Delta+1$, the algorithm recovers in expected
$O(\Delta n)$ actions. When $\mathtt{k}$ may grow arbitrarily, the algorithm
recovers in expected O(n) actions in total. Thus, our algorithm can be made
optimal with respect to space or time complexity.
| [
{
"version": "v1",
"created": "Wed, 7 May 2008 07:39:14 GMT"
},
{
"version": "v2",
"created": "Tue, 13 May 2008 08:06:10 GMT"
}
] | 2009-09-29T00:00:00 | [
[
"Bernard",
"Samuel",
"",
"LIP6"
],
[
"Devismes",
"Stéphane",
"",
"LRI"
],
[
"Potop-Butucaru",
"Maria Gradinariu",
"",
"LIP6, INRIA Rocquencourt"
],
[
"Tixeuil",
"Sébastien",
"",
"LIP6"
]
] |
0805.1071 | Zoya Svitkina | Zoya Svitkina and Lisa Fleischer | Submodular approximation: sampling-based algorithms and lower bounds | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We introduce several generalizations of classical computer science problems
obtained by replacing simpler objective functions with general submodular
functions. The new problems include submodular load balancing, which
generalizes load balancing or minimum-makespan scheduling, submodular sparsest
cut and submodular balanced cut, which generalize their respective graph cut
problems, as well as submodular function minimization with a cardinality lower
bound. We establish upper and lower bounds for the approximability of these
problems with a polynomial number of queries to a function-value oracle. The
approximation guarantees for most of our algorithms are of the order of
sqrt(n/ln n). We show that this is the inherent difficulty of the problems by
proving matching lower bounds. We also give an improved lower bound for the
problem of approximately learning a monotone submodular function. In addition,
we present an algorithm for approximately learning submodular functions with
special structure, whose guarantee is close to the lower bound. Although quite
restrictive, the class of functions with this structure includes the ones that
are used for lower bounds both by us and in previous work. This demonstrates
that if there are significantly stronger lower bounds for this problem, they
rely on more general submodular functions.
| [
{
"version": "v1",
"created": "Wed, 7 May 2008 21:37:18 GMT"
},
{
"version": "v2",
"created": "Wed, 24 Sep 2008 23:31:37 GMT"
},
{
"version": "v3",
"created": "Mon, 31 May 2010 23:57:44 GMT"
}
] | 2010-06-02T00:00:00 | [
[
"Svitkina",
"Zoya",
""
],
[
"Fleischer",
"Lisa",
""
]
] |
0805.1213 | Florin Constantin | Florin Constantin, Jon Feldman, S. Muthukrishnan and Martin Pal | Online Ad Slotting With Cancellations | 10 pages, 1 figure | null | null | null | cs.GT cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Many advertisers buy advertisements (ads) on the Internet or on traditional
media and seek simple, online mechanisms to reserve ad slots in advance. Media
publishers represent a vast and varying inventory, and they too seek automatic,
online mechanisms for pricing and allocating such reservations. In this paper,
we present and study a simple model for auctioning such ad slots in advance.
Bidders arrive sequentially and report which slots they are interested in. The
seller must decide immediately whether or not to grant a reservation. Our model
allows a seller to accept reservations, but possibly cancel the allocations
later and pay the bidder a cancellation compensation (bump payment). Our main
result is an online mechanism to derive prices and bump payments that is
efficient to implement. This mechanism has many desirable properties. It is
individually rational; winners have an incentive to be honest and bidding one's
true value dominates any lower bid. Our mechanism's efficiency is within a
constant fraction of the a posteriori optimally efficient solution. Its revenue
is within a constant fraction of the a posteriori revenue of the
Vickrey-Clarke-Groves mechanism. Our results make no assumptions about the
order of arrival of bids or the value distribution of bidders and still hold if
the items for sale are elements of a matroid, a more general setting than slot
allocation.
| [
{
"version": "v1",
"created": "Thu, 8 May 2008 18:12:50 GMT"
}
] | 2008-05-09T00:00:00 | [
[
"Constantin",
"Florin",
""
],
[
"Feldman",
"Jon",
""
],
[
"Muthukrishnan",
"S.",
""
],
[
"Pal",
"Martin",
""
]
] |
0805.1257 | Chadi Kari | Chadi Kari, Alexander Russell and Narasimha Shashidhar | Randomized Work-Competitive Scheduling for Cooperative Computing on
$k$-partite Task Graphs | null | null | null | null | cs.DC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A fundamental problem in distributed computing is the task of cooperatively
executing a given set of $t$ tasks by $p$ processors where the communication
medium is dynamic and subject to failures. The dynamics of the communication
medium lead to groups of processors being disconnected and possibly reconnected
during the entire course of the computation furthermore tasks can have
dependencies among them. In this paper, we present a randomized algorithm whose
competitive ratio is dependent on the dynamics of the communication medium and
also on the nature of the dependencies among the tasks.
| [
{
"version": "v1",
"created": "Fri, 9 May 2008 00:27:28 GMT"
},
{
"version": "v2",
"created": "Sat, 24 Mar 2012 23:52:01 GMT"
}
] | 2012-03-27T00:00:00 | [
[
"Kari",
"Chadi",
""
],
[
"Russell",
"Alexander",
""
],
[
"Shashidhar",
"Narasimha",
""
]
] |
0805.1348 | Yakov Nekrich | Marek Karpinski, Yakov Nekrich | Searching for Frequent Colors in Rectangles | null | null | null | null | cs.DS cs.CG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study a new variant of colored orthogonal range searching problem: given a
query rectangle $Q$ all colors $c$, such that at least a fraction $\tau$ of all
points in $Q$ are of color $c$, must be reported. We describe several data
structures for that problem that use pseudo-linear space and answer queries in
poly-logarithmic time.
| [
{
"version": "v1",
"created": "Fri, 9 May 2008 13:47:55 GMT"
}
] | 2008-05-12T00:00:00 | [
[
"Karpinski",
"Marek",
""
],
[
"Nekrich",
"Yakov",
""
]
] |
0805.1401 | Mustaq Ahmed | Mustaq Ahmed, Sandip Das, Sachin Lodha, Anna Lubiw, Anil Maheshwari,
Sasanka Roy | Approximation Algorithms for Shortest Descending Paths in Terrains | 24 pages, 8 figures | null | null | null | cs.CG cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A path from s to t on a polyhedral terrain is descending if the height of a
point p never increases while we move p along the path from s to t. No
efficient algorithm is known to find a shortest descending path (SDP) from s to
t in a polyhedral terrain. We give two approximation algorithms (more
precisely, FPTASs) that solve the SDP problem on general terrains. Both
algorithms are simple, robust and easy to implement.
| [
{
"version": "v1",
"created": "Fri, 9 May 2008 19:39:19 GMT"
}
] | 2008-05-12T00:00:00 | [
[
"Ahmed",
"Mustaq",
""
],
[
"Das",
"Sandip",
""
],
[
"Lodha",
"Sachin",
""
],
[
"Lubiw",
"Anna",
""
],
[
"Maheshwari",
"Anil",
""
],
[
"Roy",
"Sasanka",
""
]
] |
0805.1487 | Spyros Sioutas SS | Lagogiannis George, Lorentzos Nikos, Sioutas Spyros, Theodoridis
Evaggelos | A Time Efficient Indexing Scheme for Complex Spatiotemporal Retrieval | 6 pages, 7 figures, submitted to Sigmod Record | null | null | null | cs.DB cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The paper is concerned with the time efficient processing of spatiotemporal
predicates, i.e. spatial predicates associated with an exact temporal
constraint. A set of such predicates forms a buffer query or a Spatio-temporal
Pattern (STP) Query with time. In the more general case of an STP query, the
temporal dimension is introduced via the relative order of the spatial
predicates (STP queries with order). Therefore, the efficient processing of a
spatiotemporal predicate is crucial for the efficient implementation of more
complex queries of practical interest. We propose an extension of a known
approach, suitable for processing spatial predicates, which has been used for
the efficient manipulation of STP queries with order. The extended method is
supported by efficient indexing structures. We also provide experimental
results that show the efficiency of the technique.
| [
{
"version": "v1",
"created": "Sat, 10 May 2008 17:18:32 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"George",
"Lagogiannis",
""
],
[
"Nikos",
"Lorentzos",
""
],
[
"Spyros",
"Sioutas",
""
],
[
"Evaggelos",
"Theodoridis",
""
]
] |
0805.1598 | Peiyush Jain | Peiyush Jain | A Simple In-Place Algorithm for In-Shuffle | 3 pages | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The paper presents a simple, linear time, in-place algorithm for performing a
2-way in-shuffle which can be used with little modification for certain other
k-way shuffles.
| [
{
"version": "v1",
"created": "Mon, 12 May 2008 09:28:18 GMT"
}
] | 2008-05-13T00:00:00 | [
[
"Jain",
"Peiyush",
""
]
] |
0805.1661 | Glenn Hickey | G. Hickey, P. Carmi, A. Maheshwari, N. Zeh | NAPX: A Polynomial Time Approximation Scheme for the Noah's Ark Problem | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Noah's Ark Problem (NAP) is an NP-Hard optimization problem with
relevance to ecological conservation management. It asks to maximize the
phylogenetic diversity (PD) of a set of taxa given a fixed budget, where each
taxon is associated with a cost of conservation and a probability of
extinction. NAP has received renewed interest with the rise in availability of
genetic sequence data, allowing PD to be used as a practical measure of
biodiversity. However, only simplified instances of the problem, where one or
more parameters are fixed as constants, have as of yet been addressed in the
literature. We present NAPX, the first algorithm for the general version of NAP
that returns a $1 - \epsilon$ approximation of the optimal solution. It runs in
$O(\frac{n B^2 h^2 \log^2n}{\log^2(1 - \epsilon)})$ time where $n$ is the
number of species, and $B$ is the total budget and $h$ is the height of the
input tree. We also provide improved bounds for its expected running time.
| [
{
"version": "v1",
"created": "Mon, 12 May 2008 15:04:26 GMT"
},
{
"version": "v2",
"created": "Mon, 27 Oct 2008 18:57:31 GMT"
}
] | 2008-10-27T00:00:00 | [
[
"Hickey",
"G.",
""
],
[
"Carmi",
"P.",
""
],
[
"Maheshwari",
"A.",
""
],
[
"Zeh",
"N.",
""
]
] |
0805.2630 | Kamesh Munagala | Sudipto Guha and Kamesh Munagala | Sequential Design of Experiments via Linear Programming | The results and presentation in this paper are subsumed by the
article "Approximation algorithms for Bayesian multi-armed bandit problems"
http://arxiv.org/abs/1306.3525 | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The celebrated multi-armed bandit problem in decision theory models the basic
trade-off between exploration, or learning about the state of a system, and
exploitation, or utilizing the system. In this paper we study the variant of
the multi-armed bandit problem where the exploration phase involves costly
experiments and occurs before the exploitation phase; and where each play of an
arm during the exploration phase updates a prior belief about the arm. The
problem of finding an inexpensive exploration strategy to optimize a certain
exploitation objective is NP-Hard even when a single play reveals all
information about an arm, and all exploration steps cost the same.
We provide the first polynomial time constant-factor approximation algorithm
for this class of problems. We show that this framework also generalizes
several problems of interest studied in the context of data acquisition in
sensor networks. Our analyses also extends to switching and setup costs, and to
concave utility objectives.
Our solution approach is via a novel linear program rounding technique based
on stochastic packing. In addition to yielding exploration policies whose
performance is within a small constant factor of the adaptive optimal policy, a
nice feature of this approach is that the resulting policies explore the arms
sequentially without revisiting any arm. Sequentiality is a well-studied
concept in decision theory, and is very desirable in domains where multiple
explorations can be conducted in parallel, for instance, in the sensor network
context.
| [
{
"version": "v1",
"created": "Sat, 17 May 2008 22:48:22 GMT"
},
{
"version": "v2",
"created": "Tue, 18 Jun 2013 15:13:17 GMT"
}
] | 2013-06-19T00:00:00 | [
[
"Guha",
"Sudipto",
""
],
[
"Munagala",
"Kamesh",
""
]
] |
0805.2646 | Ilias Diakonikolas | Ilias Diakonikolas, Mihalis Yannakakis | Small Approximate Pareto Sets for Bi-objective Shortest Paths and Other
Problems | submitted full version | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We investigate the problem of computing a minimum set of solutions that
approximates within a specified accuracy $\epsilon$ the Pareto curve of a
multiobjective optimization problem. We show that for a broad class of
bi-objective problems (containing many important widely studied problems such
as shortest paths, spanning tree, and many others), we can compute in
polynomial time an $\epsilon$-Pareto set that contains at most twice as many
solutions as the minimum such set. Furthermore we show that the factor of 2 is
tight for these problems, i.e., it is NP-hard to do better. We present upper
and lower bounds for three or more objectives, as well as for the dual problem
of computing a specified number $k$ of solutions which provide a good
approximation to the Pareto curve.
| [
{
"version": "v1",
"created": "Sat, 17 May 2008 06:10:19 GMT"
}
] | 2008-05-20T00:00:00 | [
[
"Diakonikolas",
"Ilias",
""
],
[
"Yannakakis",
"Mihalis",
""
]
] |
0805.2671 | Spyros Sioutas SS | Spyros Sioutas | Finger Indexed Sets: New Approaches | 13 pages, 1 figure, Submitted to Journal of Universal Computer
Science (J.UCS) | null | null | null | cs.DS cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In the particular case we have insertions/deletions at the tail of a given
set S of $n$ one-dimensional elements, we present a simpler and more concrete
algorithm than that presented in [Anderson, 2007] achieving the same (but also
amortized) upper bound of $O(\sqrt{logd/loglogd})$ for finger searching
queries, where $d$ is the number of sorted keys between the finger element and
the target element we are looking for. Furthermore, in general case we have
insertions/deletions anywhere we present a new randomized algorithm achieving
the same expected time bounds. Even the new solutions achieve the optimal
bounds in amortized or expected case, the advantage of simplicity is of great
importance due to practical merits we gain.
| [
{
"version": "v1",
"created": "Sat, 17 May 2008 14:05:12 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Sioutas",
"Spyros",
""
]
] |
0805.2681 | Spyros Sioutas SS | Spyros Sioutas, Dimitrios Sofotassios, Kostas Tsichlas, Dimitrios
Sotiropoulos, Panayiotis Vlamos | Canonical polygon Queries on the plane: a New Approach | 7 pages, 9 figures, Accepted for publication in Journal of Computers
(JCP), http://www.informatik.uni-trier.de/~ley/db/journals/jcp/index.html | null | null | null | cs.CG cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The polygon retrieval problem on points is the problem of preprocessing a set
of $n$ points on the plane, so that given a polygon query, the subset of points
lying inside it can be reported efficiently.
It is of great interest in areas such as Computer Graphics, CAD applications,
Spatial Databases and GIS developing tasks. In this paper we study the problem
of canonical $k$-vertex polygon queries on the plane. A canonical $k$-vertex
polygon query always meets the following specific property: a point retrieval
query can be transformed into a linear number (with respect to the number of
vertices) of point retrievals for orthogonal objects such as rectangles and
triangles (throughout this work we call a triangle orthogonal iff two of its
edges are axis-parallel).
We present two new algorithms for this problem. The first one requires
$O(n\log^2{n})$ space and $O(k\frac{log^3n}{loglogn}+A)$ query time. A simple
modification scheme on first algorithm lead us to a second solution, which
consumes $O(n^2)$ space and $O(k \frac{logn}{loglogn}+A)$ query time, where $A$
denotes the size of the answer and $k$ is the number of vertices.
The best previous solution for the general polygon retrieval problem uses
$O(n^2)$ space and answers a query in $O(k\log{n}+A)$ time, where $k$ is the
number of vertices. It is also very complicated and difficult to be implemented
in a standard imperative programming language such as C or C++.
| [
{
"version": "v1",
"created": "Sat, 17 May 2008 16:00:09 GMT"
},
{
"version": "v2",
"created": "Thu, 30 Jul 2009 10:23:50 GMT"
}
] | 2009-07-30T00:00:00 | [
[
"Sioutas",
"Spyros",
""
],
[
"Sofotassios",
"Dimitrios",
""
],
[
"Tsichlas",
"Kostas",
""
],
[
"Sotiropoulos",
"Dimitrios",
""
],
[
"Vlamos",
"Panayiotis",
""
]
] |
0805.3742 | Henrik B\"a\"arnhielm | Henrik B\"a\"arnhielm | Algorithmic problems in twisted groups of Lie type | The author's PhD thesis | null | null | null | math.GR cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This thesis contains a collection of algorithms for working with the twisted
groups of Lie type known as Suzuki groups, and small and large Ree groups. The
two main problems under consideration are constructive recognition and
constructive membership testing. We also consider problems of generating and
conjugating Sylow and maximal subgroups. The algorithms are motivated by, and
form a part of, the Matrix Group Recognition Project. Obtaining both
theoretically and practically efficient algorithms has been a central goal. The
algorithms have been developed with, and implemented in, the computer algebra
system MAGMA.
| [
{
"version": "v1",
"created": "Sat, 24 May 2008 04:21:29 GMT"
},
{
"version": "v2",
"created": "Sun, 8 Jun 2008 10:34:21 GMT"
}
] | 2008-06-08T00:00:00 | [
[
"Bäärnhielm",
"Henrik",
""
]
] |
0805.3901 | Gregory Gutin | Gregory Gutin and Eun Jung Kim | Properly Coloured Cycles and Paths: Results and Open Problems | null | null | null | null | cs.DM cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we consider a number of results and seven conjectures on
properly edge-coloured (PC) paths and cycles in edge-coloured multigraphs. We
overview some known results and prove new ones. In particular, we consider a
family of transformations of an edge-coloured multigraph $G$ into an ordinary
graph that allow us to check the existence PC cycles and PC $(s,t)$-paths in
$G$ and, if they exist, to find shortest ones among them. We raise a problem of
finding the optimal transformation and consider a possible solution to the
problem.
| [
{
"version": "v1",
"created": "Mon, 26 May 2008 09:17:21 GMT"
},
{
"version": "v2",
"created": "Fri, 30 May 2008 14:33:10 GMT"
},
{
"version": "v3",
"created": "Sat, 31 May 2008 07:48:28 GMT"
}
] | 2008-05-31T00:00:00 | [
[
"Gutin",
"Gregory",
""
],
[
"Kim",
"Eun Jung",
""
]
] |
0805.4147 | Meng He | Prosenjit Bose, Eric Y. Chen, Meng He, Anil Maheshwari, Pat Morin | Succinct Geometric Indexes Supporting Point Location Queries | null | null | null | null | cs.CG cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose to design data structures called succinct geometric indexes of
negligible space (more precisely, o(n) bits) that, by taking advantage of the n
points in the data set permuted and stored elsewhere as a sequence, to support
geometric queries in optimal time. Our first and main result is a succinct
geometric index that can answer point location queries, a fundamental problem
in computational geometry, on planar triangulations in O(lg n) time. We also
design three variants of this index. The first supports point location using
$\lg n + 2\sqrt{\lg n} + O(\lg^{1/4} n)$ point-line comparisons. The second
supports point location in o(lg n) time when the coordinates are integers
bounded by U. The last variant can answer point location in O(H+1) expected
time, where H is the entropy of the query distribution. These results match the
query efficiency of previous point location structures that use O(n) words or
O(n lg n) bits, while saving drastic amounts of space.
We then generalize our succinct geometric index to planar subdivisions, and
design indexes for other types of queries. Finally, we apply our techniques to
design the first implicit data structures that support point location in
$O(\lg^2 n)$ time.
| [
{
"version": "v1",
"created": "Tue, 27 May 2008 15:15:05 GMT"
}
] | 2008-05-28T00:00:00 | [
[
"Bose",
"Prosenjit",
""
],
[
"Chen",
"Eric Y.",
""
],
[
"He",
"Meng",
""
],
[
"Maheshwari",
"Anil",
""
],
[
"Morin",
"Pat",
""
]
] |
0805.4300 | Shai Gutner | Noga Alon and Shai Gutner | Balanced Families of Perfect Hash Functions and Their Applications | null | Proc. of 34th ICALP (2007), 435-446 | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The construction of perfect hash functions is a well-studied topic. In this
paper, this concept is generalized with the following definition. We say that a
family of functions from $[n]$ to $[k]$ is a $\delta$-balanced $(n,k)$-family
of perfect hash functions if for every $S \subseteq [n]$, $|S|=k$, the number
of functions that are 1-1 on $S$ is between $T/\delta$ and $\delta T$ for some
constant $T>0$. The standard definition of a family of perfect hash functions
requires that there will be at least one function that is 1-1 on $S$, for each
$S$ of size $k$. In the new notion of balanced families, we require the number
of 1-1 functions to be almost the same (taking $\delta$ to be close to 1) for
every such $S$. Our main result is that for any constant $\delta > 1$, a
$\delta$-balanced $(n,k)$-family of perfect hash functions of size $2^{O(k \log
\log k)} \log n$ can be constructed in time $2^{O(k \log \log k)} n \log n$.
Using the technique of color-coding we can apply our explicit constructions to
devise approximation algorithms for various counting problems in graphs. In
particular, we exhibit a deterministic polynomial time algorithm for
approximating both the number of simple paths of length $k$ and the number of
simple cycles of size $k$ for any $k \leq O(\frac{\log n}{\log \log \log n})$
in a graph with $n$ vertices. The approximation is up to any fixed desirable
relative error.
| [
{
"version": "v1",
"created": "Wed, 28 May 2008 09:49:18 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Alon",
"Noga",
""
],
[
"Gutner",
"Shai",
""
]
] |
0806.0840 | Mugurel Ionut Andreica | Mugurel Ionut Andreica | A Dynamic Programming Framework for Combinatorial Optimization Problems
on Graphs with Bounded Pathwidth | Some of the ideas presented in this paper were later used by the
author for preparing algorithmic tasks for several contests where the author
was a member of the scientific committee (e.g. ACM ICPC Southeastern regional
contest 2009 and Balkan olympiad in informatics 2011). Such tasks (including
the task statement and solutions) can be found in the attached zip archive;
THETA 16 / AQTR, Cluj-Napoca : Romania (2008) | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we present an algorithmic framework for solving a class of
combinatorial optimization problems on graphs with bounded pathwidth. The
problems are NP-hard in general, but solvable in linear time on this type of
graphs. The problems are relevant for assessing network reliability and
improving the network's performance and fault tolerance. The main technique
considered in this paper is dynamic programming.
| [
{
"version": "v1",
"created": "Wed, 4 Jun 2008 19:18:45 GMT"
},
{
"version": "v2",
"created": "Mon, 17 Dec 2012 14:03:58 GMT"
}
] | 2012-12-18T00:00:00 | [
[
"Andreica",
"Mugurel Ionut",
""
]
] |
0806.0928 | Martin N\"ollenburg | Martin N\"ollenburg, Danny Holten, Markus V\"olker, Alexander Wolff | Drawing Binary Tanglegrams: An Experimental Evaluation | see
http://www.siam.org/proceedings/alenex/2009/alx09_011_nollenburgm.pdf | Proceedings of the 11th Workshop on Algorithm Engineering and
Experiments (ALENEX'09), pages 106-119. SIAM, April 2009 | null | null | cs.DS cs.CG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A binary tanglegram is a pair <S,T> of binary trees whose leaf sets are in
one-to-one correspondence; matching leaves are connected by inter-tree edges.
For applications, for example in phylogenetics or software engineering, it is
required that the individual trees are drawn crossing-free. A natural
optimization problem, denoted tanglegram layout problem, is thus to minimize
the number of crossings between inter-tree edges.
The tanglegram layout problem is NP-hard and is currently considered both in
application domains and theory. In this paper we present an experimental
comparison of a recursive algorithm of Buchin et al., our variant of their
algorithm, the algorithm hierarchy sort of Holten and van Wijk, and an integer
quadratic program that yields optimal solutions.
| [
{
"version": "v1",
"created": "Thu, 5 Jun 2008 11:00:33 GMT"
}
] | 2009-05-15T00:00:00 | [
[
"Nöllenburg",
"Martin",
""
],
[
"Holten",
"Danny",
""
],
[
"Völker",
"Markus",
""
],
[
"Wolff",
"Alexander",
""
]
] |
0806.0983 | Sandy Irani | Joan Boyar, Sandy Irani, Kim S. Larsen | A Comparison of Performance Measures for Online Algorithms | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper provides a systematic study of several proposed measures for
online algorithms in the context of a specific problem, namely, the two server
problem on three colinear points. Even though the problem is simple, it
encapsulates a core challenge in online algorithms which is to balance
greediness and adaptability. We examine Competitive Analysis, the Max/Max
Ratio, the Random Order Ratio, Bijective Analysis and Relative Worst Order
Analysis, and determine how these measures compare the Greedy Algorithm, Double
Coverage, and Lazy Double Coverage, commonly studied algorithms in the context
of server problems. We find that by the Max/Max Ratio and Bijective Analysis,
Greedy is the best of the three algorithms. Under the other measures, Double
Coverage and Lazy Double Coverage are better, though Relative Worst Order
Analysis indicates that Greedy is sometimes better. Only Bijective Analysis and
Relative Worst Order Analysis indicate that Lazy Double Coverage is better than
Double Coverage. Our results also provide the first proof of optimality of an
algorithm under Relative Worst Order Analysis.
| [
{
"version": "v1",
"created": "Thu, 5 Jun 2008 14:50:08 GMT"
},
{
"version": "v2",
"created": "Fri, 12 Oct 2012 16:03:13 GMT"
}
] | 2012-10-15T00:00:00 | [
[
"Boyar",
"Joan",
""
],
[
"Irani",
"Sandy",
""
],
[
"Larsen",
"Kim S.",
""
]
] |
0806.1722 | Guangwu Xu | George Davida, Bruce Litow and Guangwu Xu | Fast Arithmetics Using Chinese Remaindering | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, some issues concerning the Chinese remaindering representation
are discussed. Some new converting methods, including an efficient
probabilistic algorithm based on a recent result of von zur Gathen and
Shparlinski \cite{Gathen-Shparlinski}, are described. An efficient refinement
of the NC$^1$ division algorithm of Chiu, Davida and Litow
\cite{Chiu-Davida-Litow} is given, where the number of moduli is reduced by a
factor of $\log n$.
| [
{
"version": "v1",
"created": "Tue, 10 Jun 2008 18:21:09 GMT"
}
] | 2008-06-11T00:00:00 | [
[
"Davida",
"George",
""
],
[
"Litow",
"Bruce",
""
],
[
"Xu",
"Guangwu",
""
]
] |
0806.1948 | Kai-Min Chung | Kai-Min Chung, Salil Vadhan | Tight Bounds for Hashing Block Sources | An extended abstract of this paper will appear in RANDOM08 | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | It is known that if a 2-universal hash function $H$ is applied to elements of
a {\em block source} $(X_1,...,X_T)$, where each item $X_i$ has enough
min-entropy conditioned on the previous items, then the output distribution
$(H,H(X_1),...,H(X_T))$ will be ``close'' to the uniform distribution. We
provide improved bounds on how much min-entropy per item is required for this
to hold, both when we ask that the output be close to uniform in statistical
distance and when we only ask that it be statistically close to a distribution
with small collision probability. In both cases, we reduce the dependence of
the min-entropy on the number $T$ of items from $2\log T$ in previous work to
$\log T$, which we show to be optimal. This leads to corresponding improvements
to the recent results of Mitzenmacher and Vadhan (SODA `08) on the analysis of
hashing-based algorithms and data structures when the data items come from a
block source.
| [
{
"version": "v1",
"created": "Wed, 11 Jun 2008 19:54:14 GMT"
}
] | 2008-06-12T00:00:00 | [
[
"Chung",
"Kai-Min",
""
],
[
"Vadhan",
"Salil",
""
]
] |
0806.1978 | Luca Trevisan | Luca Trevisan | Max Cut and the Smallest Eigenvalue | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We describe a new approximation algorithm for Max Cut. Our algorithm runs in
$\tilde O(n^2)$ time, where $n$ is the number of vertices, and achieves an
approximation ratio of $.531$. On instances in which an optimal solution cuts a
$1-\epsilon$ fraction of edges, our algorithm finds a solution that cuts a
$1-4\sqrt{\epsilon} + 8\epsilon-o(1)$ fraction of edges.
Our main result is a variant of spectral partitioning, which can be
implemented in nearly linear time. Given a graph in which the Max Cut optimum
is a $1-\epsilon$ fraction of edges, our spectral partitioning algorithm finds
a set $S$ of vertices and a bipartition $L,R=S-L$ of $S$ such that at least a
$1-O(\sqrt \epsilon)$ fraction of the edges incident on $S$ have one endpoint
in $L$ and one endpoint in $R$. (This can be seen as an analog of Cheeger's
inequality for the smallest eigenvalue of the adjacency matrix of a graph.)
Iterating this procedure yields the approximation results stated above.
A different, more complicated, variant of spectral partitioning leads to an
$\tilde O(n^3)$ time algorithm that cuts $1/2 + e^{-\Omega(1/\eps)}$ fraction
of edges in graphs in which the optimum is $1/2 + \epsilon$.
| [
{
"version": "v1",
"created": "Thu, 12 Jun 2008 17:51:02 GMT"
},
{
"version": "v2",
"created": "Sun, 15 Jun 2008 05:09:08 GMT"
},
{
"version": "v3",
"created": "Mon, 22 Sep 2008 23:59:20 GMT"
},
{
"version": "v4",
"created": "Wed, 24 Sep 2008 09:39:48 GMT"
},
{
"version": "v5",
"created": "Mon, 8 Dec 2008 19:03:46 GMT"
}
] | 2008-12-08T00:00:00 | [
[
"Trevisan",
"Luca",
""
]
] |
0806.2068 | Fran\c{c}ois Nicolas | Francois Nicolas | A simple, polynomial-time algorithm for the matrix torsion problem | 6 pages. Not intended to be submitted | null | null | null | cs.DM cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Matrix Torsion Problem (MTP) is: given a square matrix M with rational
entries, decide whether two distinct powers of M are equal. It has been shown
by Cassaigne and the author that the MTP reduces to the Matrix Power Problem
(MPP) in polynomial time: given two square matrices A and B with rational
entries, the MTP is to decide whether B is a power of A. Since the MPP is
decidable in polynomial time, it is also the case of the MTP. However, the
algorithm for MPP is highly non-trivial. The aim of this note is to present a
simple, direct, polynomial-time algorithm for the MTP.
| [
{
"version": "v1",
"created": "Thu, 12 Jun 2008 13:24:46 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Jul 2008 23:51:08 GMT"
},
{
"version": "v3",
"created": "Tue, 8 Sep 2009 19:46:11 GMT"
}
] | 2009-09-08T00:00:00 | [
[
"Nicolas",
"Francois",
""
]
] |
0806.2274 | Marko A. Rodriguez | Marko A. Rodriguez, Joshua Shinavier | Exposing Multi-Relational Networks to Single-Relational Network Analysis
Algorithms | ISSN:1751-1577 | Journal of Informetrics, volume 4, number 1, pages 29-41, 2009 | 10.1016/j.joi.2009.06.004 | LA-UR-08-03931 | cs.DM cs.DS | http://creativecommons.org/licenses/publicdomain/ | Many, if not most network analysis algorithms have been designed specifically
for single-relational networks; that is, networks in which all edges are of the
same type. For example, edges may either represent "friendship," "kinship," or
"collaboration," but not all of them together. In contrast, a multi-relational
network is a network with a heterogeneous set of edge labels which can
represent relationships of various types in a single data structure. While
multi-relational networks are more expressive in terms of the variety of
relationships they can capture, there is a need for a general framework for
transferring the many single-relational network analysis algorithms to the
multi-relational domain. It is not sufficient to execute a single-relational
network analysis algorithm on a multi-relational network by simply ignoring
edge labels. This article presents an algebra for mapping multi-relational
networks to single-relational networks, thereby exposing them to
single-relational network analysis algorithms.
| [
{
"version": "v1",
"created": "Fri, 13 Jun 2008 16:07:19 GMT"
},
{
"version": "v2",
"created": "Wed, 9 Dec 2009 16:08:02 GMT"
}
] | 2009-12-09T00:00:00 | [
[
"Rodriguez",
"Marko A.",
""
],
[
"Shinavier",
"Joshua",
""
]
] |
0806.2287 | Shiva Kasiviswanathan | Martin Furer and Shiva Prasad Kasiviswanathan | Approximately Counting Embeddings into Random Graphs | Earlier version appeared in Random 2008. Fixed an typo in Definition
3.1 | Combinator. Probab. Comp. 23 (2014) 1028-1056 | 10.1017/S0963548314000339 | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Let H be a graph, and let C_H(G) be the number of (subgraph isomorphic)
copies of H contained in a graph G. We investigate the fundamental problem of
estimating C_H(G). Previous results cover only a few specific instances of this
general problem, for example, the case when H has degree at most one
(monomer-dimer problem). In this paper, we present the first general subcase of
the subgraph isomorphism counting problem which is almost always efficiently
approximable. The results rely on a new graph decomposition technique.
Informally, the decomposition is a labeling of the vertices such that every
edge is between vertices with different labels and for every vertex all
neighbors with a higher label have identical labels. The labeling implicitly
generates a sequence of bipartite graphs which permits us to break the problem
of counting embeddings of large subgraphs into that of counting embeddings of
small subgraphs. Using this method, we present a simple randomized algorithm
for the counting problem. For all decomposable graphs H and all graphs G, the
algorithm is an unbiased estimator. Furthermore, for all graphs H having a
decomposition where each of the bipartite graphs generated is small and almost
all graphs G, the algorithm is a fully polynomial randomized approximation
scheme.
We show that the graph classes of H for which we obtain a fully polynomial
randomized approximation scheme for almost all G includes graphs of degree at
most two, bounded-degree forests, bounded-length grid graphs, subdivision of
bounded-degree graphs, and major subclasses of outerplanar graphs,
series-parallel graphs and planar graphs, whereas unbounded-length grid graphs
are excluded.
| [
{
"version": "v1",
"created": "Fri, 13 Jun 2008 17:06:01 GMT"
},
{
"version": "v2",
"created": "Fri, 21 Jun 2013 18:40:00 GMT"
}
] | 2019-02-20T00:00:00 | [
[
"Furer",
"Martin",
""
],
[
"Kasiviswanathan",
"Shiva Prasad",
""
]
] |
0806.2707 | Pat Morin | Vida Dujmovic, John Howat, and Pat Morin | Biased Range Trees | null | null | null | null | cs.CG cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A data structure, called a biased range tree, is presented that preprocesses
a set S of n points in R^2 and a query distribution D for 2-sided orthogonal
range counting queries. The expected query time for this data structure, when
queries are drawn according to D, matches, to within a constant factor, that of
the optimal decision tree for S and D. The memory and preprocessing
requirements of the data structure are O(n log n).
| [
{
"version": "v1",
"created": "Tue, 17 Jun 2008 15:18:40 GMT"
}
] | 2008-06-18T00:00:00 | [
[
"Dujmovic",
"Vida",
""
],
[
"Howat",
"John",
""
],
[
"Morin",
"Pat",
""
]
] |
0806.3201 | Gueorgi Kossinets | Gueorgi Kossinets, Jon Kleinberg, Duncan Watts | The Structure of Information Pathways in a Social Communication Network | 9 pages, 10 figures, to appear in Proceedings of the 14th ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining (KDD'08),
August 24-27, 2008, Las Vegas, Nevada, USA | null | null | null | physics.soc-ph cs.DS physics.data-an | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Social networks are of interest to researchers in part because they are
thought to mediate the flow of information in communities and organizations.
Here we study the temporal dynamics of communication using on-line data,
including e-mail communication among the faculty and staff of a large
university over a two-year period. We formulate a temporal notion of "distance"
in the underlying social network by measuring the minimum time required for
information to spread from one node to another -- a concept that draws on the
notion of vector-clocks from the study of distributed computing systems. We
find that such temporal measures provide structural insights that are not
apparent from analyses of the pure social network topology. In particular, we
define the network backbone to be the subgraph consisting of edges on which
information has the potential to flow the quickest. We find that the backbone
is a sparse graph with a concentration of both highly embedded edges and
long-range bridges -- a finding that sheds new light on the relationship
between tie strength and connectivity in social networks.
| [
{
"version": "v1",
"created": "Thu, 19 Jun 2008 14:22:25 GMT"
}
] | 2008-06-20T00:00:00 | [
[
"Kossinets",
"Gueorgi",
""
],
[
"Kleinberg",
"Jon",
""
],
[
"Watts",
"Duncan",
""
]
] |
0806.3258 | Daniel Karapetyan | Gregory Gutin, Daniel Karapetyan | Local Search Heuristics For The Multidimensional Assignment Problem | 30 pages. A preliminary version is published in volume 5420 of
Lecture Notes Comp. Sci., pages 100-115, 2009 | Journal of Heuristics 17(3) (2011), 201--249 | 10.1007/s10732-010-9133-3 | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Multidimensional Assignment Problem (MAP) (abbreviated s-AP in the case
of s dimensions) is an extension of the well-known assignment problem. The most
studied case of MAP is 3-AP, though the problems with larger values of s also
have a large number of applications. We consider several known neighborhoods,
generalize them and propose some new ones. The heuristics are evaluated both
theoretically and experimentally and dominating algorithms are selected. We
also demonstrate a combination of two neighborhoods may yield a heuristics
which is superior to both of its components.
| [
{
"version": "v1",
"created": "Thu, 19 Jun 2008 18:31:51 GMT"
},
{
"version": "v2",
"created": "Sat, 12 Jul 2008 22:18:26 GMT"
},
{
"version": "v3",
"created": "Fri, 5 Sep 2008 19:57:41 GMT"
},
{
"version": "v4",
"created": "Mon, 20 Oct 2008 05:41:16 GMT"
},
{
"version": "v5",
"created": "Tue, 14 Apr 2009 14:22:13 GMT"
},
{
"version": "v6",
"created": "Sat, 25 Jul 2009 11:25:38 GMT"
}
] | 2015-02-24T00:00:00 | [
[
"Gutin",
"Gregory",
""
],
[
"Karapetyan",
"Daniel",
""
]
] |
0806.3301 | Ryan Tibshirani | Ryan J. Tibshirani | Fast computation of the median by successive binning | 14 pages, 1 Postscript figure | null | null | null | stat.CO cs.DS stat.AP | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper describes a new median algorithm and a median approximation
algorithm. The former has O(n) average running time and the latter has O(n)
worst-case running time. These algorithms are highly competitive with the
standard algorithm when computing the median of a single data set, but are
significantly faster in updating the median when more data is added.
| [
{
"version": "v1",
"created": "Fri, 20 Jun 2008 00:44:53 GMT"
},
{
"version": "v2",
"created": "Tue, 12 May 2009 04:46:56 GMT"
}
] | 2009-05-12T00:00:00 | [
[
"Tibshirani",
"Ryan J.",
""
]
] |
0806.3437 | Hang Dinh | Hang Dinh and Alexander Russell | Quantum and Randomized Lower Bounds for Local Search on
Vertex-Transitive Graphs | null | null | null | null | quant-ph cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study the problem of \emph{local search} on a graph. Given a real-valued
black-box function f on the graph's vertices, this is the problem of
determining a local minimum of f--a vertex v for which f(v) is no more than f
evaluated at any of v's neighbors. In 1983, Aldous gave the first strong lower
bounds for the problem, showing that any randomized algorithm requires
$\Omega(2^{n/2 - o(1)})$ queries to determine a local minima on the
n-dimensional hypercube. The next major step forward was not until 2004 when
Aaronson, introducing a new method for query complexity bounds, both
strengthened this lower bound to $\Omega(2^{n/2}/n^2)$ and gave an analogous
lower bound on the quantum query complexity. While these bounds are very
strong, they are known only for narrow families of graphs (hypercubes and
grids). We show how to generalize Aaronson's techniques in order to give
randomized (and quantum) lower bounds on the query complexity of local search
for the family of vertex-transitive graphs. In particular, we show that for any
vertex-transitive graph G of N vertices and diameter d, the randomized and
quantum query complexities for local search on G are $\Omega(N^{1/2}/d\log N)$
and $\Omega(N^{1/4}/\sqrt{d\log N})$, respectively.
| [
{
"version": "v1",
"created": "Fri, 20 Jun 2008 18:46:50 GMT"
}
] | 2008-06-23T00:00:00 | [
[
"Dinh",
"Hang",
""
],
[
"Russell",
"Alexander",
""
]
] |
0806.3471 | Maria Gradinariu Potop-Butucaru | Davide Canepa and Maria Gradinariu Potop-Butucaru | Stabilizing Tiny Interaction Protocols | null | null | null | null | cs.DC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we present the self-stabilizing implementation of a class of
token based algorithms. In the current work we only consider interactions
between weak nodes. They are uniform, they do not have unique identifiers, are
static and their interactions are restricted to a subset of nodes called
neighbours. While interacting, a pair of neighbouring nodes may create mobile
agents (that materialize in the current work the token abstraction) that
perform traversals of the network and accelerate the system stabilization. In
this work we only explore the power of oblivious stateless agents.
Our work shows that the agent paradigm is an elegant distributed tool for
achieving self-stabilization in Tiny Interaction Protocols (TIP). Nevertheless,
in order to reach the full power of classical self-stabilizing algorithms more
complex classes of agents have to be considered (e.g. agents with memory,
identifiers or communication skills). Interestingly, our work proposes for the
first time a model that unifies the recent studies in mobile robots(agents)
that evolve in a discrete space and the already established population
protocols paradigm.
| [
{
"version": "v1",
"created": "Fri, 20 Jun 2008 21:01:52 GMT"
},
{
"version": "v2",
"created": "Tue, 15 Jun 2010 11:39:09 GMT"
}
] | 2010-06-16T00:00:00 | [
[
"Canepa",
"Davide",
""
],
[
"Potop-Butucaru",
"Maria Gradinariu",
""
]
] |
0806.3668 | Bodo Manthey | Markus Bl\"aser, Bodo Manthey, Oliver Putz | Approximating Multi-Criteria Max-TSP | An extended abstract of this worl will appear in Proc. of the 16th
Ann. European Symposium on Algorithms (ESA 2008) | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present randomized approximation algorithms for multi-criteria Max-TSP.
For Max-STSP with k > 1 objective functions, we obtain an approximation ratio
of $1/k - \eps$ for arbitrarily small $\eps > 0$. For Max-ATSP with k objective
functions, we obtain an approximation ratio of $1/(k+1) - \eps$.
| [
{
"version": "v1",
"created": "Mon, 23 Jun 2008 12:28:10 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Bläser",
"Markus",
""
],
[
"Manthey",
"Bodo",
""
],
[
"Putz",
"Oliver",
""
]
] |
0806.3827 | Mugurel Ionut Andreica | Mugurel Ionut Andreica | Optimal Scheduling of File Transfers with Divisible Sizes on Multiple
Disjoint Paths | The algorithmic techniques presented in this paper (particularly the
block partitioning framework) were used as part of the official solutions for
several tasks proposed by the author in the 2012 Romanian National Olympiad
in Informatics (the statements and solutions for these tasks can be found in
the attached zip archive) | Proceedings of the IEEE Romania International Conference
"Communications", 2008. (ISBN: 978-606-521-008-0), Bucharest : Romania (2008) | null | null | cs.DS cs.NI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper I investigate several offline and online data transfer
scheduling problems and propose efficient algorithms and techniques for
addressing them. In the offline case, I present a novel, heuristic, algorithm
for scheduling files with divisible sizes on multiple disjoint paths, in order
to maximize the total profit (the problem is equivalent to the multiple
knapsack problem with divisible item sizes). I then consider a cost
optimization problem for transferring a sequence of identical files, subject to
time constraints imposed by the data transfer providers. For the online case I
propose an algorithmic framework based on the block partitioning method, which
can speed up the process of resource allocation and reservation.
| [
{
"version": "v1",
"created": "Tue, 24 Jun 2008 07:16:26 GMT"
},
{
"version": "v2",
"created": "Thu, 20 Dec 2012 08:42:41 GMT"
}
] | 2012-12-21T00:00:00 | [
[
"Andreica",
"Mugurel Ionut",
""
]
] |
0806.4073 | Frank Gurski | Frank Gurski | A comparison of two approaches for polynomial time algorithms computing
basic graph parameters | 25 pages, 3 figures | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we compare and illustrate the algorithmic use of graphs of
bounded tree-width and graphs of bounded clique-width. For this purpose we give
polynomial time algorithms for computing the four basic graph parameters
independence number, clique number, chromatic number, and clique covering
number on a given tree structure of graphs of bounded tree-width and graphs of
bounded clique-width in polynomial time. We also present linear time algorithms
for computing the latter four basic graph parameters on trees, i.e. graphs of
tree-width 1, and on co-graphs, i.e. graphs of clique-width at most 2.
| [
{
"version": "v1",
"created": "Wed, 25 Jun 2008 11:26:47 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Gurski",
"Frank",
""
]
] |
0806.4361 | Yakov Nekrich | Marek Karpinski, Yakov Nekrich | Space Efficient Multi-Dimensional Range Reporting | null | null | null | null | cs.DS cs.CG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present a data structure that supports three-dimensional range reporting
queries in $O(\log \log U + (\log \log n)^3+k)$ time and uses $O(n\log^{1+\eps}
n)$ space, where $U$ is the size of the universe, $k$ is the number of points
in the answer,and $\eps$ is an arbitrary constant. This result improves over
the data structure of Alstrup, Brodal, and Rauhe (FOCS 2000) that uses
$O(n\log^{1+\eps} n)$ space and supports queries in $O(\log n+k)$ time,the data
structure of Nekrich (SoCG'07) that uses $O(n\log^{3} n)$ space and supports
queries in $O(\log \log U + (\log \log n)^2 + k)$ time, and the data structure
of Afshani (ESA'08) that uses $O(n\log^{3} n)$ space and also supports queries
in $O(\log \log U + (\log \log n)^2 + k)$ time but relies on randomization
during the preprocessing stage. Our result allows us to significantly reduce
the space usage of the fastest previously known static and incremental
$d$-dimensional data structures, $d\geq 3$, at a cost of increasing the query
time by a negligible $O(\log \log n)$ factor.
| [
{
"version": "v1",
"created": "Thu, 26 Jun 2008 16:32:57 GMT"
},
{
"version": "v2",
"created": "Fri, 24 Apr 2009 12:08:08 GMT"
}
] | 2009-04-24T00:00:00 | [
[
"Karpinski",
"Marek",
""
],
[
"Nekrich",
"Yakov",
""
]
] |
0806.4372 | Stavros Nikolopoulos D. | Katerina Asdre and Stavros D. Nikolopoulos | The 1-fixed-endpoint Path Cover Problem is Polynomial on Interval Graph | null | null | null | null | cs.DS cs.DM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider a variant of the path cover problem, namely, the
$k$-fixed-endpoint path cover problem, or kPC for short, on interval graphs.
Given a graph $G$ and a subset $\mathcal{T}$ of $k$ vertices of $V(G)$, a
$k$-fixed-endpoint path cover of $G$ with respect to $\mathcal{T}$ is a set of
vertex-disjoint paths $\mathcal{P}$ that covers the vertices of $G$ such that
the $k$ vertices of $\mathcal{T}$ are all endpoints of the paths in
$\mathcal{P}$. The kPC problem is to find a $k$-fixed-endpoint path cover of
$G$ of minimum cardinality; note that, if $\mathcal{T}$ is empty the stated
problem coincides with the classical path cover problem. In this paper, we
study the 1-fixed-endpoint path cover problem on interval graphs, or 1PC for
short, generalizing the 1HP problem which has been proved to be NP-complete
even for small classes of graphs. Motivated by a work of Damaschke, where he
left both 1HP and 2HP problems open for the class of interval graphs, we show
that the 1PC problem can be solved in polynomial time on the class of interval
graphs. The proposed algorithm is simple, runs in $O(n^2)$ time, requires
linear space, and also enables us to solve the 1HP problem on interval graphs
within the same time and space complexity.
| [
{
"version": "v1",
"created": "Thu, 26 Jun 2008 18:13:31 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Asdre",
"Katerina",
""
],
[
"Nikolopoulos",
"Stavros D.",
""
]
] |
0806.4652 | Yong Gao | Yong Gao | A Fixed-Parameter Algorithm for Random Instances of Weighted d-CNF
Satisfiability | 13 pages | null | null | null | cs.DS cs.AI cs.CC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study random instances of the weighted $d$-CNF satisfiability problem
(WEIGHTED $d$-SAT), a generic W[1]-complete problem. A random instance of the
problem consists of a fixed parameter $k$ and a random $d$-CNF formula
$\weicnf{n}{p}{k, d}$ generated as follows: for each subset of $d$ variables
and with probability $p$, a clause over the $d$ variables is selected uniformly
at random from among the $2^d - 1$ clauses that contain at least one negated
literals.
We show that random instances of WEIGHTED $d$-SAT can be solved in $O(k^2n +
n^{O(1)})$-time with high probability, indicating that typical instances of
WEIGHTED $d$-SAT under this instance distribution are fixed-parameter
tractable. The result also hold for random instances from the model
$\weicnf{n}{p}{k,d}(d')$ where clauses containing less than $d' (1 < d' < d)$
negated literals are forbidden, and for random instances of the renormalized
(miniaturized) version of WEIGHTED $d$-SAT in certain range of the random
model's parameter $p(n)$. This, together with our previous results on the
threshold behavior and the resolution complexity of unsatisfiable instances of
$\weicnf{n}{p}{k, d}$, provides an almost complete characterization of the
typical-case behavior of random instances of WEIGHTED $d$-SAT.
| [
{
"version": "v1",
"created": "Sat, 28 Jun 2008 05:03:47 GMT"
}
] | 2008-12-18T00:00:00 | [
[
"Gao",
"Yong",
""
]
] |