Quantitative invertibility of random matrices: a combinatorial perspective

Quantitative invertibility of random matrices: a combinatorial perspective, Discrete Analysis 2021:10, 38 pp. This paper concerns the general and much studied question of how likely a large random real or complex square matrix is to be invertible, and, if it is invertible, how far it is from being...

Full description

Bibliographic Details
Main Author: Vishesh Jain
Format: Article
Language:English
Published: Diamond Open Access Journals 2021-09-01
Series:Discrete Analysis
Online Access:https://discrete-analysis.scholasticahq.com/article/27994-quantitative-invertibility-of-random-matrices-a-combinatorial-perspective.pdf
id doaj-febdf7c844964ba5a20eb8d9227f29b7
record_format Article
collection DOAJ
language English
format Article
sources DOAJ
author Vishesh Jain
spellingShingle Vishesh Jain
Quantitative invertibility of random matrices: a combinatorial perspective
Discrete Analysis
author_facet Vishesh Jain
author_sort Vishesh Jain
title Quantitative invertibility of random matrices: a combinatorial perspective
title_short Quantitative invertibility of random matrices: a combinatorial perspective
title_full Quantitative invertibility of random matrices: a combinatorial perspective
title_fullStr Quantitative invertibility of random matrices: a combinatorial perspective
title_full_unstemmed Quantitative invertibility of random matrices: a combinatorial perspective
title_sort quantitative invertibility of random matrices: a combinatorial perspective
publisher Diamond Open Access Journals
series Discrete Analysis
issn 2397-3129
publishDate 2021-09-01
description Quantitative invertibility of random matrices: a combinatorial perspective, Discrete Analysis 2021:10, 38 pp. This paper concerns the general and much studied question of how likely a large random real or complex square matrix is to be invertible, and, if it is invertible, how far it is from being singular. To interpret the latter question, note that an $n\times n$ matrix $A$ is singular if and only if there is a vector $x$, belonging to $\mathbb R^n$ or $\mathbb C^n$ as appropriate, such that $\|x\|_2=1$ and $\|Ax\|_2=0$. Thus, a natural quantitative measure of the invertibility of a matrix is the smallest $\epsilon$ such that there exists $x$ with $\|x\|_2=1$ and $\|Ax\|_2=\epsilon$. This is equal to the smallest singular value of $A$, denoted $s_n(A)$, and it is also the reciprocal of the operator norm of $A^{-1}$. The most famous question in this general area is a problem of Komlós that is deceptively simple to state: what is the probability that an $n\times n$ matrix with independent random entries in $\{-1,1\}$ is singular? It is widely believed that the main reason that such a matrix might fail to be invertible is that it has two rows or two columns that are either equal or that add up to zero. The probability of this is approximately $2\binom n22^{-(n-1)}$, which is approximately $2n^22^{-n}$. Even obtaining an $o(1)$ bound was non-trivial, and achieved first by Komlós. There followed a long series of improvements: the first exponentially small bound was obtained by Kahn, Komlós and Szemerédi, much improved exponents were obtained by Tao and Vu, and by Bourgain, Vu and Wood, and in a recent breakthrough of Tikhomirov a bound of $(1/2+o(1))^n$ was obtained, so that we now know the correct exponent. The papers of Tao and Vu introduced into the area an important set of ideas known as inverse Littlewood-Offord theory. The Littlewood-Offord problem asks the following: given a set of vectors $v_1,\dots,v_n$, how large can the probability be that a random sum $\sum_i\epsilon_iv_i$ is equal to some particular vector $v$, where $\epsilon_1,\dots,\epsilon_n$ are chosen independently at random from $\{-1,1\}$? More generally, one can ask similar questions with the $\epsilon_i$ being other real or complex random variables that satisfy suitable conditions. One can also ask for the probability that the sum belongs to some small ball rather than equalling a particular vector -- for instance, this is clearly the right question when the $\epsilon_i$ are continuous random variables. An _inverse_ Littlewood-Offord problem seeks to understand the structure of the set $\{v_1,\dots,v_n\}$ when the above probability is large. When proving that a random matrix is unlikely to be singular, one often ends up in a situation where one would like to apply a union bound, but a naive union bound fails because of a few bad cases. These cases turn out to correspond to sets of vectors where the Littlewood-Offord probability is large. However, if one can show that such sets have strong structural properties, then one can also show that they are rare, and in that way split up the union bound into a small "structured part" and a much larger "typical part", where in each part the naive union bound works -- in one case because one is adding up only a very small number of largish probabilities, and in the other case because one is adding up a large number of very small probabilities. It has subsequently been understood, starting with work of Rudelson and Vershynin, that for many purposes the inverse Littlewood-Offord theory is in a certain sense "too strong" a tool to use. In principle, to obtain an upper bound on the number of bad cases, one does not need to describe the structure of those cases, which is hard to do, especially if one wants tight bounds -- one just needs to show that they are rare. In a paper of Ferber, Jain, Luh and Samotij it was shown that it is possible to bypass the use of inverse Littlewood-Offord theory by using clever counting arguments instead, and thereby to obtain better bounds for certain problems. This paper continues this line of thought, looking at a random matrix model $M_n=M+N_n$, where $M$ is a fixed $n\times n$ complex matrix and $N_n$ is a random matrix with independent entries of mean 0 and variance 1. This model arises naturally in computer science, where $M$ can be thought of as the matrix that an algorithm would like to work with, and $N_n$ is a bit of noise that is in practice added to it. It was previously known, using inverse Littlewood-Offord theory and assuming a power-type upper bound on $\|M\|$, that for every positive constant $A$ there exists a positive constant $B$ such that the probability that $s_n(M_n)\leq n^{-B}$ is at most $n^{-A}$. The main result of this paper extends the range of applicability of this result so that $\|M\|$ is allowed to be larger, and the bounds $n^{-A}$ and $n^{-B}$ are replaced by bounds of the form $\exp(-n^c)$. Such results with exponential bounds were already known in the case $M=0$, using more geometric techniques originally due to Rudelson and Vershynin and further developed by them and other authors. However, for matrices with non-centred entries the results are new.
url https://discrete-analysis.scholasticahq.com/article/27994-quantitative-invertibility-of-random-matrices-a-combinatorial-perspective.pdf
work_keys_str_mv AT visheshjain quantitativeinvertibilityofrandommatricesacombinatorialperspective
_version_ 1717779891698532352
spelling doaj-febdf7c844964ba5a20eb8d9227f29b72021-09-06T07:43:58ZengDiamond Open Access JournalsDiscrete Analysis2397-31292021-09-01Quantitative invertibility of random matrices: a combinatorial perspectiveVishesh JainQuantitative invertibility of random matrices: a combinatorial perspective, Discrete Analysis 2021:10, 38 pp. This paper concerns the general and much studied question of how likely a large random real or complex square matrix is to be invertible, and, if it is invertible, how far it is from being singular. To interpret the latter question, note that an $n\times n$ matrix $A$ is singular if and only if there is a vector $x$, belonging to $\mathbb R^n$ or $\mathbb C^n$ as appropriate, such that $\|x\|_2=1$ and $\|Ax\|_2=0$. Thus, a natural quantitative measure of the invertibility of a matrix is the smallest $\epsilon$ such that there exists $x$ with $\|x\|_2=1$ and $\|Ax\|_2=\epsilon$. This is equal to the smallest singular value of $A$, denoted $s_n(A)$, and it is also the reciprocal of the operator norm of $A^{-1}$. The most famous question in this general area is a problem of Komlós that is deceptively simple to state: what is the probability that an $n\times n$ matrix with independent random entries in $\{-1,1\}$ is singular? It is widely believed that the main reason that such a matrix might fail to be invertible is that it has two rows or two columns that are either equal or that add up to zero. The probability of this is approximately $2\binom n22^{-(n-1)}$, which is approximately $2n^22^{-n}$. Even obtaining an $o(1)$ bound was non-trivial, and achieved first by Komlós. There followed a long series of improvements: the first exponentially small bound was obtained by Kahn, Komlós and Szemerédi, much improved exponents were obtained by Tao and Vu, and by Bourgain, Vu and Wood, and in a recent breakthrough of Tikhomirov a bound of $(1/2+o(1))^n$ was obtained, so that we now know the correct exponent. The papers of Tao and Vu introduced into the area an important set of ideas known as inverse Littlewood-Offord theory. The Littlewood-Offord problem asks the following: given a set of vectors $v_1,\dots,v_n$, how large can the probability be that a random sum $\sum_i\epsilon_iv_i$ is equal to some particular vector $v$, where $\epsilon_1,\dots,\epsilon_n$ are chosen independently at random from $\{-1,1\}$? More generally, one can ask similar questions with the $\epsilon_i$ being other real or complex random variables that satisfy suitable conditions. One can also ask for the probability that the sum belongs to some small ball rather than equalling a particular vector -- for instance, this is clearly the right question when the $\epsilon_i$ are continuous random variables. An _inverse_ Littlewood-Offord problem seeks to understand the structure of the set $\{v_1,\dots,v_n\}$ when the above probability is large. When proving that a random matrix is unlikely to be singular, one often ends up in a situation where one would like to apply a union bound, but a naive union bound fails because of a few bad cases. These cases turn out to correspond to sets of vectors where the Littlewood-Offord probability is large. However, if one can show that such sets have strong structural properties, then one can also show that they are rare, and in that way split up the union bound into a small "structured part" and a much larger "typical part", where in each part the naive union bound works -- in one case because one is adding up only a very small number of largish probabilities, and in the other case because one is adding up a large number of very small probabilities. It has subsequently been understood, starting with work of Rudelson and Vershynin, that for many purposes the inverse Littlewood-Offord theory is in a certain sense "too strong" a tool to use. In principle, to obtain an upper bound on the number of bad cases, one does not need to describe the structure of those cases, which is hard to do, especially if one wants tight bounds -- one just needs to show that they are rare. In a paper of Ferber, Jain, Luh and Samotij it was shown that it is possible to bypass the use of inverse Littlewood-Offord theory by using clever counting arguments instead, and thereby to obtain better bounds for certain problems. This paper continues this line of thought, looking at a random matrix model $M_n=M+N_n$, where $M$ is a fixed $n\times n$ complex matrix and $N_n$ is a random matrix with independent entries of mean 0 and variance 1. This model arises naturally in computer science, where $M$ can be thought of as the matrix that an algorithm would like to work with, and $N_n$ is a bit of noise that is in practice added to it. It was previously known, using inverse Littlewood-Offord theory and assuming a power-type upper bound on $\|M\|$, that for every positive constant $A$ there exists a positive constant $B$ such that the probability that $s_n(M_n)\leq n^{-B}$ is at most $n^{-A}$. The main result of this paper extends the range of applicability of this result so that $\|M\|$ is allowed to be larger, and the bounds $n^{-A}$ and $n^{-B}$ are replaced by bounds of the form $\exp(-n^c)$. Such results with exponential bounds were already known in the case $M=0$, using more geometric techniques originally due to Rudelson and Vershynin and further developed by them and other authors. However, for matrices with non-centred entries the results are new.https://discrete-analysis.scholasticahq.com/article/27994-quantitative-invertibility-of-random-matrices-a-combinatorial-perspective.pdf