May 21

Add to Calendar 2024-05-21 16:15:00 2024-05-21 17:15:00 America/New_York Circuit Complexity from Circuit Analysis Algorithms Abstract: I will explain how I got into circuit complexity, and how I ended up proving theorems like "NEXP does not have small ACC circuits". (The answer may surprise you! Or maybe not.) I'll also talk about what else has been proved, building on the ideas that were introduced. 32-G449

May 14

Add to Calendar 2024-05-14 16:15:00 2024-05-14 17:15:00 America/New_York Low-Step Multi-Commodity Flow Emulators Abstract: We introduce the concept of low-step multi-commodity flow emulators for any undirected, capacitated graph. At a high level, these emulators contain approximate multi-commodity flows whose paths contain a small number of edges, shattering the infamous flow decomposition barrier for multi-commodity flow.We prove the existence of low-step multi-commodity flow emulators and develop efficient algorithms to compute them. We then apply them to solve constant-approximate $k$-commodity flow in $O((m+k)^{1+\epsilon})$ time. To bypass the $O(mk)$ flow decomposition barrier, we represent our output multi-commodity flow implicitly; prior to our work, even the existence of implicit constant-approximate multi-commodity flows of size $o(mk)$ was unknown.Our results generalize to the \emph{minimum cost} setting, where each edge has an associated cost and the multi-commodity flow must satisfy a cost budget. Our algorithms are also parallel. 32-G449

May 07

Add to Calendar 2024-05-07 16:15:00 2024-05-07 17:15:00 America/New_York Parallel Derandomization for Chernoff-like Concentrations Abstract: Randomized algorithms frequently use concentration results such as Chernoff and Hoeffding bounds. A longstanding challenge in parallel computing is to devise an efficient method to derandomize parallel algorithms that rely on such concentrations. Classic works of Motwani, Naor, and Naor [FOCS'89] and Berger and Rompel [FOCS'89] provide an elegant parallel derandomization method for these, via a binary search in a k-wise independent space, but with one major disadvanage: it blows up the computational work by a (large) polynomial. That is, the resulting deterministic parallel algorithm is far from work-efficiency and needs polynomial processors even to match the speed of single-processor computation. This talk overviews a duo of recent papers that provide the first nearly work-efficient parallel derandomization method for Chernoff-like concentrations.Based on joint work with Christoph Grunau and Vaclav Rozhon, which can be accessed via https://arxiv.org/abs/2311.13764 and https://arxiv.org/abs/2311.13771. 32-G882

May 01

Add to Calendar 2024-05-01 11:00:00 2024-05-01 13:00:00 America/New_York Efficient Algorithms for Vector Similarities Title: Efficient Algorithms for Vector SimilaritiesAbstract:A key cog in machine learning is the humble embedding: a vector representation of real world objects such as text, images, graphs, and even molecules. It is common to curate massive datasets of embeddings by inferencing on a ML model of choice. However, the sheer dataset size and large dimensionality is often *the* bottleneck in effectively leveraging and learning from this rich dataset. Inspired by this computational bottleneck in modern machine learning pipelines, we study the following question: "How can we efficiently compute on large scale high dimensional data?"In this thesis, we focus on two aspects of this question.(1) Fast local similarity computation: we give faster algorithms to compute complicated notions of similarity, such as those between collections of vectors (think optimal transport), as well as dimensionality reduction techniques which preserve similarities. In addition to computational efficiency, other resource constraints such as space and privacy are also considered.(2) Fast global similairty analysis: we study algorithms for analyzing global relationships between vectors encoded in similarity matrices. Our algorithms compute on similarity matrices, such as distance or kernel matrices, without ever initializing them, thus avoiding a quadratic time and space bottleneck.Overall, my main message is that sublinear algorithms design principles are instrumental in designing scalable algorithms for big data. For the presentation, I will only cover a subset of results in each of the categories, with a big emphasis on simplicity.Thesis Committee:Piotr Indyk (Advisor) - MITRonitt Rubinfeld - MITHuy Nguyen - Northeastern University Seminar Room G882 (Hewlett Room)

April 23

Add to Calendar 2024-04-23 16:15:00 2024-04-23 17:15:00 America/New_York On some Recent Dynamic Graph Algorithms Abstract: We discuss some recent algorithms for problems in dynamic graphs undergoing edge insertions and deletions. In the first part of the talk, we will discuss connections between the approximate bipartite matching problem in fully dynamic graphs, and versions of the online matrix-vector multiplication conjecture. These connections will lead to faster algorithms in online and offline settings, as well as some conditional lower bounds.In the second part of the talk, we will briefly discuss how interior point methods can be used to design algorithms for partially dynamic graphs: those undergoing only edge insertions (incremental) or only edge deletions (decremental). This leads to almost-optimal algorithms for problems including incremental cycle detection, and decremental s-t distance. 32-G882

April 16

Add to Calendar 2024-04-16 16:15:00 2024-04-16 17:15:00 America/New_York Succinct arguments for QMA from compiled nonlocal games Abstract: We construct a succinct classical argument system for QMA, the quantum analogue of NP, from generic and standard cryptographic assumptions. Previously, building on the prior work of Mahadev (FOCS '18), Bartusek et al. (CRYPTO '22) also constructed a succinct classical argument system for QMA. However, their construction relied on post-quantumly secure indistinguishability obfuscation, a very strong primitive which is not known from standard cryptographic assumptions. In contrast, the primitives we use (namely, collapsing hash functions and a mild version of quantum homomorphic encryption) are much weaker and are implied by standard assumptions such as LWE. Our protocol is constructed using a general transformation which was designed by Kalai et al. (STOC '23) as a candidate method to compile any quantum nonlocal game into an argument system. Our main technical contribution is to analyze the soundness of this transformation when it is applied to a succinct self-test for Pauli measurements on maximally entangled states, the latter of which is a key component in the proof of MIP* = RE in quantum complexity.Joint work with Tony Metger (ETH Zurich) and Tina Zhang (MIT) 32-G449

March 19

Add to Calendar 2024-03-19 16:15:00 2024-03-19 17:15:00 America/New_York From Large to Small Datasets: Size Generalization for Clustering Algorithm Selection Abstract: In clustering algorithm selection, we are given a massive dataset and must efficiently select which clustering algorithm to use. We study this problem in a semi-supervised setting, with an unknown ground-truth clustering that we can only access through expensive oracle queries. Ideally, the clustering algorithm's output will be structurally close to the ground truth. We approach this problem by introducing a notion of size generalization for clustering algorithm accuracy. We identify conditions under which we can (1) subsample the massive clustering instance, (2) evaluate a set of candidate algorithms on the smaller instance, and (3) guarantee that the algorithm with the best accuracy on the small instance will have the best accuracy on the original big instance. We verify these findings both theoretically and empirically.This is joint work with Vaggos Chatziafratis and Ishani Karmarkar. 32-G449

February 27

Add to Calendar 2024-02-27 16:15:00 2024-02-27 17:15:00 America/New_York Towards a Complexity Theory for the Quantum Age Abstract: How hard is it to compress a quantum state? To fast-forward the evolution of a local Hamiltonian? To unscramble the Hawking radiation of a black hole? Traditional complexity theory -- which is centered around decision problems and tasks with classical inputs and outputs -- appears inadequate for reasoning about the complexity of such tasks involving quantum inputs and outputs.I'll discuss why we need a "fully quantum" complexity theory, and will describe some facets of such a theory. As a key illustration I'll explain how a "fully quantum" task called the Uhlmann Transformation Problem characterizes the complexity of seemingly unrelated problems, such as decoding noisy quantum channels, performing the Harlow-Hayden black hole radiation decoding task, and breaking the security of quantum bit commitment schemes. I will describe some of the many open problems and directions to explore in the world of fully quantum complexity theory.Bio: Henry Yuen is an Associate Professor of Computer Science at Columbia University. His research focuses on the interplay between quantum computing, complexity theory, cryptography, and information theory. Yuen received a BA in mathematics from the University of Southern California in 2010, and received his PhD in computer science at MIT in 2016. He is a recipient of the NSF CAREER award and a Sloan Fellowship. 32-G449

February 20

Add to Calendar 2024-02-20 16:15:00 2024-02-20 17:15:00 America/New_York The Subspace Flatness Conjecture and Faster Integer Programming Abstract (in latex):In a seminal paper, Kannan and Lov\'asz (1988) considered a quantity $\mu_{KL}(\Lambda,K)$which denotes the best volume-based lower bound on the \emph{covering radius} $\mu(\Lambda,K)$ of a convexbody $K$ with respect to a lattice $\Lambda$. Kannan and Lov\'asz proved that $\mu(\Lambda,K) \leq n \cdot \mu_{KL}(\Lambda,K)$ and the Subspace Flatness Conjecture by Dadush (2012) claims a $O(\log n)$ factor suffices, which would matchthe lower bound from the work of Kannan and Lov\'asz.We settle this conjecture up to a constant in the exponent by proving that $\mu(\Lambda,K) \leq O(\log^{3}(n)) \cdot \mu_{KL} (\Lambda,K)$. Our proof isbased on the Reverse Minkowski Theorem due to Regev and Stephens-Davidowitz (2017).Following the work of Dadush (2012, 2019), we obtain a $(\log n)^{O(n)}$-time randomized algorithm tosolve integer programs in $n$ variables.Another implication of our main result is a near-optimal \emph{flatness constant} of $O(n \log^{3}(n))$.This is joint work with Victor Reis. 32-G449

February 06

Add to Calendar 2024-02-06 16:15:00 2024-02-06 17:00:00 America/New_York Global Convergence of Shifted QR Abstract: The Shifted QR algorithm is the most widely used algorithm since the 1960's for computing the eigenvalues and eigenvectors of a dense matrix. It is a bit like the simplex algorithm in that it is specified by a "shifting strategy" (cf. pivoting strategies for simplex). There are shifting strategies which are typically very efficient in practice and on special classes of matrices, but occasionally fail, and no strategy is known to be provably rapidly convergent on every input matrix. I will report on some recent significant progress on this question, in particular a new shifting rule which converges rapidly on a small random perturbation of every matrix. Joint work with Jorge Garza Vargas and Jess Banks. 32-G449