Thursday, September 8 |
07:00 - 09:00 |
Breakfast (Vistas Dining Room) |
09:00 - 09:50 |
Ran Raz: Fast Learning Requires Good Memory: A Time-Space Lower Bound for Parity Learning ↓ We prove that any algorithm for learning parities requires either a memory of quadratic size or an exponential number of samples. This proves a recent conjecture of Steinhardt, Valiant and Wager and shows that for some learning problems a large storage space is crucial.
More formally, in the problem of parity learning, an unknown string x∈{0,1}n was chosen uniformly at random. A learner tries to learn x from a stream of samples (a1,b1),(a2,b2)..., where each at is uniformly distributed over {0,1}n and bt is the inner product of at and x, modulo 2. We show that any algorithm for parity learning, that uses less than n2/25 bits of memory, requires an exponential number of samples.
Previously, there was no non-trivial lower bound on the number of samples needed, for any learning problem, even if the allowed memory size is O(n) (where n is the space needed to store one sample).
We also give an application of our result in the field of bounded-storage cryptography. We show an encryption scheme that requires a private key of length n, as well as time complexity of n per encryption/decryption of each bit, and is provenly and unconditionally secure as long as the attacker uses less than n2/25 memory bits and the scheme is used at most an exponential number of times. Previous works on bounded-storage cryptography assumed that the memory size used by the attacker is at most linear in the time needed for encryption/decryption. (TCPL 201) |
09:50 - 10:10 |
Coffee Break (TCPL Foyer) |
10:10 - 11:00 |
Marco Carmosino: Learning algorithms from natural proofs ↓ Circuit analysis algorithms such as learning, SAT, minimum circuit size, and compression imply circuit lower bounds. A natural question is: given that these algorithms are sufficient for circuit lower bounds, to what degree are they necessary? We make progress on this question by showing a generic implication in the opposite direction: from natural proofs of circuit lower bounds (in the sense of Razborov and Rudich) to randomized learning and compression algorithms. This is the first such implication outside of the derandomization setting. As an application, we use known natural lower bounds for AC0[p] circuits (due to Razborov and Smolensky) to get the first quasi-polynomial time algorithm for learning AC0[p] functions, in the PAC model over the uniform distribution, with membership queries. (TCPL 201) |
11:05 - 11:55 |
Rahul Santhanam: Stronger Connections between Learning, Lower Bounds and Pseudorandomness ↓ We prove several results giving new and stronger connections between
learning theory, circuit complexity and pseudorandomness, including:
1. (Learning Speedups) Let C be any "typical" class of Boolean circuits, and
C[s(n)] denote n-variable C-circuits of size at most s(n). If C[poly] admits a
randomized weak learning algorithm under the uniform distribution with
membership queries that runs in time 2^n/n^{\omega(1)}, then for every k and
\varepsilon > 0 the class C[n^k] can be learned to high accuracy in time
O(2^{n^\varepsilon}).
2.(Equivalences) There is \varepsilon > 0 such that C[2^{n^{\varepsilon}}] can
be learned in time 2^n/n^{\omega(1)} if and only if C[poly] can be learned in
time 2^{O((\log n)^c)}. Furthermore, our proof technique yields equivalences
between various randomized learning models in the exponential and
sub-exponential time settings, such as learning with membership queries,
learning with membership and equivalence queries, and the related framework of
truth-table compression.
3.(Learning versus Pseudorandom Functions) In the non-uniform setting, there is
non-trivial learning for C[poly(n)] if and only if there are no secure
pseudorandom function families computable in C[poly].
4.(Lower Bounds from Nontrivial Learning Algorithms) Let C be any class of
Boolean circuits closed under projection. If for each k, C[n^k] admits a
randomized weak learning algorithm with membership queries under the uniform
distribution that runs in time 2^n/n^{\omega(1)}, then for each k, BPE is not
contained in C[n^k]. If there are P-natural proofs useful against
C[2^{n^{o(1)}}], then ZPEXP is not contained in C[poly].
Among the proof techniques we use are the uniform hardness-randomness tradeoffs
of [Impagliazzo-Wigderson, Trevisan-Vadhan], the easy witness method of
[Kabanets, Impagliazzo-Kabanets-Wigderson], the interpretation of the
Nisan-Wigderson generator as a pseudo-random function generator due to
[Carmosino-Impagliazzo-Kabanets-Kolokolova] and the approximate min-max theorem
of [Althofer, Lipton-Young].
This is joint work with Igor Carboni Oliveira. (TCPL 201) |
12:00 - 13:00 |
Lunch (Vistas Dining Room) |
13:10 - 14:00 |
Daniel Kane: A New Approach to Distribution Testing ↓ We study the problem of determining whether or not a
discrete distribution has a given property from a small number of
samples. We present a new technique in this field that operates by
reducing many such problems in a black box way to a simple L^2 tester.
We show how this new technique can recover simple proofs of several
known, optimal results and how it can be extended to provide optimal
solutions to a number of previously unsolved problems. (TCPL 201) |
14:05 - 14:55 |
Shachar Lovett: Structure of protocols for XOR functions ↓ Let f be a boolean function on n variables. Its associated XOR function is the two-party function F(x, y) = f(x xor y). If f has a parity decision tree of depth d, then it gives a natural deterministic protocol for F which sends 2d bits. We establish the reverse direction: any deterministic protocol for F which sends c bits, implies a parity decision tree for f of depth poly(c). The proof combines a novel technique of entropy reduction for protocols, with existing techniques in Fourier analysis and additive combinatorics.
Joint work with Hamed Hatami and Kaave Hosseini. (TCPL 201) |
15:00 - 15:30 |
Coffee Break (TCPL Foyer) |
15:30 - 15:55 |
Anindya De: Learning "Sums of independent commonly supported integer random variables" ↓ Let S be a fixed subset of natural numbers and consider the family of random variables which can be generated as sums of independent random variables each supported on S (abbreviated as SICSIRVs on S). So, for example, if S = {0,1}, then the SICSIRVs on S is the class of sums of independent Bernoulli random variables (aka Poisson Binomial random variables).
Recent work by Daskalakis et. al. (STOC 12, FOCS 13) has shown that if S is a subset of [-k, .. , k], then the complexity of learning SICSIRVs on S is poly(k/\epsilon) (and hence independent of n, the number of summands!). We investigate the problem of learning SICSIRVs on S for arbitrary S. Our results are as follows:
(a) If |S| <=3, then the complexity of learning SICSIRV on S is poly(1/\epsilon) (and independent of S).
(b) There exists S ={a_1, a_2, a_3, a_4} such that the complexity of learning SICSIRV on S is \Omega(log log max a_i) and this is tight.
Thus, our results show a sharp transition in the complexity of learning of SICSIRVs when the size of S goes from 3 to 4. The upper bounds are based on a new central limit theorem while the lower bound exploits delicate constructions based on continued fractions.
Joint work with Phil Long and Rocco Servedio. (TCPL 201) |
16:00 - 16:50 |
Oded Regev: The Reverse Minkowski Theorem ↓ Informally, Minkowski's first theorem states that lattices that are globally dense (have small determinant) are also locally dense (have lots of points in a small ball around the origin). This fundamental result dates back to 1891 and has a very wide range of applications.
Here we prove a reverse form of the Minkowski's theorem, showing that locally dense lattice are also globally dense (in the appropriate sense).
We also discuss the many applications of this result. In particular, it resolves a conjecture by Saloff-Coste on the behavior of random walks on flat tori, has implications to the complexity of lattice problems, and also makes progress on conjectures by Kannan and Lovasz, and by Shapira and Weiss that are closely related to a celebrated conjecture by Minkowski.
Based on joint works with Daniel Dadush and Noah Stephens-Davidowitz. (TCPL 201) |
17:30 - 19:30 |
Dinner (Vistas Dining Room) |
20:00 - 20:50 |
Scott Aaronson: Complexity-Theoretic Foundations of Quantum Supremacy Experiments ↓ Depending on time, I'll discuss some subset of the following results.
(1) There exists an oracle relative to which classical and quantum
computers can solve the same approximate sampling problems in
polynomial time, and yet the polynomial hierarchy is infinite. This
means that any "BosonSampling-type theorem" will need to be
non-relativizing.
(2) Can P be separated from BQP by an efficiently-computable oracle
(one in P/poly)? Yes if quantum-resistant one-way functions exist; no
if P=PSPACE. The latter builds on recent results in quantum query
complexity by myself and Shalev Ben-David.
(3) How hard is it to approximately sample the output distribution of
a random quantum circuit? One can do it (and even calculate the
probabilities) in polynomial space and m^O(n) time, where m is the
number of gates and n is the number of qubits. But if that's
essentially optimal for estimating probabilities, then approximate
sampling also requires exponential time.
Joint work with Lijie Chen (Tsinghua University) (TCPL 201) |