December notes
18 Dec 2018Karchmer-Wigderson theorem
Karchmer-Wigderson theorem makes a precise link between circuit and communication complexity. More precisely it relates the depth of a boolean circuit computing some function $f$, to the communication complexity of a game based on $f$. The game is the following:
- Inputs: Alice has an input $x$ such that $f(x)=1$, and Bob has another input $y$ such that $f(y)=0$.
- Task: find $i$ such that $x_i \neq y_i$ (such an $i$ must exist).
The way to go from the circuit to a communication protocol is the following. Alice and Bob each run the circuit on their inputs, the usual bottom-up way. Thus each of them knows the evaluation after each gate (“after this gate, this wire holds a 0, after this one it’s a 1” etc.). Now the protocol will go top-down. As after the final gate (that is, the top gate), Alice gets a 1 and Bob gets a 0, we know that the two wires entering this gate cannot both have the same value for Alice and Bob. Alice and Bob can communicate to agree on one wire that has different value. This takes constant number of bits. And then you just follow the wire to the next gate and continue. At the end Alice and Bob agree on an input wire that has different value and they win the game. The communication complexity is the circuit depth.
For the other direction, see the Internet.
(I learnt about Karchmer-Wigderson theorem recently by watching parts of an online talk by Mika Göös.)
Counting crows
These days I often cross the Jardin des Plantes in Paris, and I saw a sign giving information about the counting of crows in Paris. Basically a large number of crows now have a ring with a specific number. The problem is that the rings tend to fall, disappear etc. To evaluate this loss rate, the birds are rung with two rings. Then by counting the number of crows with zero, one and two rings, one can evaluate the loss rate, and then the population.
This sounded very much like CS to me, for example in networks, sending packets that may disappear, and trying to know the message loss for example.
Minimum degree spanning tree
Minimum-degree spanning tree, consists in finding a spanning tree of a graph, with the constraint that the maximum degree of the tree should be as small as possible. This problem is NP-hard, because if the answer is 2 then it means that you have an Hamiltonian path. But one can get an approximation with additive constant 1. The algorithm (from this paper, explained in these lecture notes) consists in a local search, with swapping of edges to as local moves.
This problems looks very natural and potentially very useful. I’d be curious to know if there are real-world applications.
(I discovered this problem in this distributed computing paper.)
Cycle cover
An (edge) cycle-cover is a set of cycles of a graph, such that every edge of the graph is contained in at least one cycle. There are several problems associated with this object, for example related to the total length of the cycles in such a cover. I want to mention a super-cute-super-hard open problem: the cycle double cover conjecture. The conjecture is: in every graph with no bridge, there is a list of cycles so that every edge is contained in exactly two.
If you think “How can this be open?!”, I’ll just add that it is listed in the open problem garden as of “outstanding importance”. See there, or on the wikipedia page linked above for the details.
(A paper about cycle covers recently appeared on the arxiv, reminded me this problem.)
MST algorithms
Consider the following algorithm for finding a minimum weight spanning tree, similar to Borůvka’s algorithm.
- Every node begins has a so-called fragment.
- Until there is only one fragment: choose an arbitrary fragment, find the lightest edge linking a node of this fragment to another fragment, merge the two fragments into a new one by adding this edge.
- Output the final fragment.
This algorithm is actually a generalization of Borůvka, Prim and Kruskal algorithm! For Borůvka’s algorithm, one would basically choose an outgoing edge for all fragments in parallel. For Prim, one would always choose the same fragment to enlarge. F or Kruskal, one would choose the lightest outgoing edge of all the fragments.
This can probably be explained with red-blue rules, but I like this point of view with a kind of scheduler, with different strategies.
(I’m working again on some minimum spanning tree distributed problems, and it reminded me about this fact, that I discovered a few years ago, but that I’ve never seen in undergraduate courses.)
Squashed cube conjecture and distance labelling
Distance labelling are labels given to the nodes of a graph such that given the labels of two arbitrary nodes $u$ and $v$, and without seeing the graph, one can compute the distance between $u$ and $v$. One usually tries to minimize the size of the labels. A strategy for this is to find some metric embedding of the graph, because then one can simply give the coordinates of the nodes as labels. In this direction, a natural thing to do is to try the embed the graph in an hypercube with the Hamming distance. This cannot be done in general, but the squashed cube conjecture (that is actually a theorem) provides a result close to this.
Consider that instead of two symbols, there are three: 0, 1, and *, and that the “distance” between $x$ and $y$ is the number of indices such that $x=0$ and $y=1$ or $x=1$ and $y=0$. That is * is at distance zero from 0 and 1. (Note that this distance is not a proper distance.) Now how many symbols do you need to have a distance preserving embedding? Exactly $n-1$, as proved in this paper.
(I discovered this in the related works section of this paper.)
Jaccard metric
Lipton and Regan talk about the Jaccard metric in a post of their blog, Gödel’s lost letter, in particular why it is a metric. I didn’t know about this distance over sets, so here is the definition.
Let $A$ and $B$ be two sets, the distance is: \(d(A,B)=1-\frac{|A \cap B|}{|A \cup B|}.\)
Preprints
I have two new preprints this month:
- Graph classes and forbidden patterns on three vertices with Michel Habib.
- Lower bounds for text indexing with mismatches and differences with Vincent Cohen-Addad and Tatiana Starikoskaya.
I really like both papers, I hope I’ll find some time to blog about it soon.