# Query-Efficient Algorithm to Find all Nash Equilibria in a Two-Player Zero-Sum Matrix Game
**Authors**: Arnab Maiti, Ross Boczar, Kevin Jamieson, Lillian J. Ratliff, University of Washington
## Abstract
We study the query complexity of finding the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ in two-player zero-sum matrix games. Fearnley and Savani [17] showed that for any randomized algorithm, there exists an $n\times n$ input matrix where it needs to query $\Omega(n^{2})$ entries in expectation to compute a single Nash equilibrium. On the other hand, Bienstock et al. [5] showed that there is a special class of matrices for which one can query $O(n)$ entries and compute its set of all Nash equilibria. However, these results do not fully characterize the query complexity of finding the set of all Nash equilibria in two-player zero-sum matrix games.
In this work, we characterize the query complexity of finding the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ in terms of the number of rows $n$ of the input matrix $A\in\mathbb{R}^{n\times n}$ , row support size $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ , and column support size $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ . We design a simple yet non-trivial randomized algorithm that, with probability $1-\delta$ , returns the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ by querying at most $O(nk^{5}\cdot\text{polylog}(n/\delta))$ entries of the input matrix $A\in\mathbb{R}^{n\times n}$ , where $k:=\max\{k_{1},k_{2}\}$ . This upper bound is tight up to a factor of $\text{poly}(k)$ , as we show that for any randomized algorithm, there exists an $n\times n$ input matrix with $\min\{k_{1},k_{2}\}=1$ , for which it needs to query $\Omega(nk)$ entries in expectation in order to find the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ .
## 1 Introduction
Von Neumann’s celebrated minimax theorem [22] initiated the study of two-player zero-sum matrix games by demonstrating the existence of a Nash equilibrium. Subsequently, it was established that two-player zero-sum matrix games are equivalent to linear programming [12, 1, 8]. With the development of polynomial-time algorithms for linear programming, it became evident that one can compute a Nash equilibrium for a two-player zero-sum matrix game in polynomial time. This implication has motivated key works at the intersection of game theory and theoretical computer science [9, 13, 14, 15, 4].
One of these works was the development of query-efficient algorithms for finding equilibrium solutions. In the context of two-player zero-sum games, a surprising result was demonstrated by Bienstock et al. [5]. They presented a deterministic algorithm that finds the strict saddle point (if it exists) by querying $O(n)$ entries of the input matrix $A\in\mathbb{R}^{n\times n}$ . A strict saddle point is an entry of a matrix that is uniquely the smallest in its row and uniquely the largest in its column. In game theory, such an entry is referred to as a strict Pure Strategy Nash equilibrium. In contrast, Fearnley and Savani [17] showed that for any randomized algorithm, there exists an $n\times n$ input matrix $A\in\mathbb{R}^{n\times n}$ from which the algorithm needs to query $\Omega(n^{2})$ entries in expectation to compute a single Nash equilibrium. This lower bound arises when the Nash equilibrium is a mixed strategy with all rows and columns in the support of this mixed strategy. The results of Bienstock et al. [5] and Fearnley and Savani [17] represent two extremes on the spectrum of solution sizes, highlighting a significant gap in the existing literature. Recently, Auger et al. [2] attempted to address this gap by studying a class of matrices with unique Nash equilibrium and binary entries, characterizing the query complexity in terms of the support size of the Nash equilibrium strategy. However, such a characterization for arbitrary matrices in $\mathbb{R}^{n\times n}$ that may have multiple Nash equilibria remains unaddressed in prior work. This leads us to the following question:
Can we characterize the query complexity of finding the set of all Nash equilibria of a matrix $A\in\mathbb{R}^{n\times n}$ in terms of an appropriate notion of the Nash equilibria solution size?
In this work, we address the above question. To this end, let us briefly review two-player zero-sum games on an input matrix $A\in\mathbb{R}^{n\times n}$ . Let $\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}$ denote the $(n-1)$ -dimensional probability simplex. A pair $(x_{\star},y_{\star})\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}\times\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}$ is a Nash equilibrium of an input matrix $A\in\mathbb{R}^{n\times n}$ if and only if the following holds for all $(x,y)\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44 444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}\times \mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}$ :
$$
\langle x,Ay_{\star}\rangle\leq\langle x_{\star},Ay_{\star}\rangle\leq\langle x
_{\star},Ay\rangle. \tag{1}
$$
Von Neumann’s minimax theorem states that $\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41 112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\min_{y\in \mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle= \min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41 112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{x\in \mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ (denoted as the value of the game $V_{A}^{\star}$ ). Thus, $(x_{\star},y_{\star})$ is a Nash equilibrium of the matrix $A$ if and only if
$$
x_{\star}\in\mathcal{X}_{\star}:=\arg\max_{x\in\mathchoice{\includegraphics[he
ight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle\quad\text{and}\quad y
_{\star}\in\mathcal{Y}_{\star}:=\arg\min_{y\in\mathchoice{\includegraphics[hei
ght=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle.
$$
An important property of the sets $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ is that they are convex, closed, and bounded polytopes (see [6]). In any zero-sum game, there is either exactly one Nash equilibrium $(x_{\star},y_{\star})$ or infinitely many. However, the value $\langle x_{\star},Ay_{\star}\rangle$ is unique and equal to $V_{A}^{\star}$ . We refer the reader to Section 2 for additional properties of Nash equilibrium, as well as for details on how to efficiently represent the sets $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ .
Our work diverges from prior work as we focus on finding the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . This allows us to perform a full strategic analysis of a game, whereas finding only a single equilibrium provides merely a partial view. For example, finding the set of all Nash equilibria allows us to identify all non-degenerate parts of the game that contain equilibrium solutions. Such a non-degenerate part is typically a square submatrix with a unique Nash equilibrium. Additionally, finding the set of all Nash equilibria helps us determine which rows (resp. columns) are suboptimal, in the sense that these rows (resp. columns) are not best response strategies to some Nash equilibrium strategy of the column (resp. row) player.
In the rest of this section, we outline the appropriate notion of Nash equilibria solution size, detail our contributions and techniques, discuss related works, and introduce important notations.
### 1.1 Appropriate notion of the Nash equilibria solution size
Recall that $V_{A}^{\star}:=\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt ,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height =3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim= 0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.219 4pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[hei ght=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}}\langle x,Ay\rangle=\min_{y\in\mathchoice{\includegraphics[h eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[h eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ , $\mathcal{X}_{\star}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ , and $\mathcal{Y}_{\star}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ . This work aims to characterize the query complexity of finding the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ in terms of the dimension $n$ of the input matrix $A\in\mathbb{R}^{n\times n}$ and the support sizes $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ and $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ . To motivate why these are appropriate notions of Nash equilibria solution size, we first present the following important result by Bohnenblust et al. [7].
**Lemma 1 (Bohnenblust et al. [7])**
*Consider an input matrix $A\in\mathbb{R}^{n\times n}$ . Let $I_{x}=\{j\in[n]:V_{A}^{\star}=\langle x,Ae_{j}\rangle\}$ and $I_{y}=\{i\in[n]:V_{A}^{\star}=\langle e_{i},Ay\rangle\}$ , where $e_{i}$ denotes the vector with a $1$ in the $i$ -th coordinate and $0 0$ ’s elsewhere. For any $v\in\mathbb{R}^{n}$ , let $\operatorname*{supp}(v):=\{i\in[n]\mid v_{i}\neq 0\}$ . Then we have the following:
$$
\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)=\bigcap\limits
_{y\in\mathcal{Y}_{\star}}I_{y}\quad\text{and}\quad\bigcup\limits_{y\in
\mathcal{Y}_{\star}}\operatorname*{supp}(y)=\bigcap\limits_{x\in\mathcal{X}_{
\star}}I_{x}
$$*
The above lemma suggests that rows not in $\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ are strictly worse than the rows in $\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ . This is because for all $y^{*}\in\mathcal{Y}^{\star}$ and $i\in\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ , we have $V_{A}^{\star}=\langle e_{i},Ay^{*}\rangle$ . In contrast, there exists a $\hat{y}\in\mathcal{Y}^{\star}$ such that for all $i\notin\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ , $V_{A}^{\star}>\langle e_{i},A\hat{y}\rangle$ . Hence, the number of optimal rows $|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ is an appropriate measure of the solution size for the row player. Using similar logic, we can justify that the number of optimal columns $|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ is an appropriate measure of the solution size for the column player.
Moreover, it is easy to observe that changing the value of the entries $(i,j)$ where $i\notin\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ and $j\notin\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)$ does not affect the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . This implies that only $O(n\cdot\max\{k_{1},k_{2}\})$ elements are important in determining the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ , which further highlights why $k_{1}$ and $k_{2}$ are appropriate notions of the solution size.
### 1.2 Our contributions and innovative techniques
Let $k:=\max\{k_{1},k_{2}\}$ , where $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ and $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ . Our main technical contribution in this paper is the following upper bound result.
**Theorem 1**
*There exists a randomized algorithm that, with probability at least $1-\delta$ , queries $O(nk^{5}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of the input matrix $A\in\mathbb{R}^{n\times n}$ and returns the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ .*
We outline the innovative techniques used to achieve the result above. As discussed in the previous section, the results of Bohnenblust et al. [7] suggest a separation between the sub-matrix with row indices $\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ and column indices $\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)$ (which we will refer to as the optimal sub-matrix) and the rest of the matrix. Our goal is to formalize this separation and therefore leverage it to identify the set of all Nash equilibria while minimizing the number of queries required. In the specific case of strict Pure Strategy Nash Equilibrium (strict PSNE), this separation is inherently present, as a strict PSNE is uniquely the least in its row and uniquely the highest in its column. Therefore, one possible approach involves constructing a large matrix with ${n\choose k_{1}}$ rows and ${n\choose k_{2}}$ columns, where each entry is bijectively mapped to a $k_{1}\times k_{2}$ sub-matrix of $A$ . The values of these entries are determined by their corresponding sub-matrices in $A$ , and this large matrix possesses a strict PSNE that maps to the optimal sub-matrix of $A$ .
Assuming we can construct such a matrix, it remains unclear how to query the entries of this large matrix without significantly increasing the query complexity of our algorithm. The algorithm proposed by Bienstock et al. [5] is inadequate for this purpose because it queries entries diagonally. If we applied a similar diagonal querying strategy to our large matrix, we would end up querying all entries of the original matrix $A$ . Instead, we propose a randomized algorithm for finding strict PSNE that operates differently from the one described by Bienstock et al. [5]. Specifically, if we have access to query oracles where a single call returns either the entire row or the entire column, our algorithm can identify the strict PSNE by making $O(\log^{2}(r))$ oracle calls, where $r$ is the total number of rows and columns in the input matrix with a strict PSNE. We can also design a query oracle for the larger matrix that returns a row or column by querying at most $O(nk)$ entries of $A$ . Hence, the query complexity of finding the strict PSNE of the large matrix will be nearly $n\cdot\text{poly}(k)$ .
One approach to creating such a large matrix is to compute the value of the game on the sub-matrix mapped to each entry and use that as the entry’s value. While this method ensures that the entry in the resulting matrix that maps to the optimal sub-matrix is a Pure Strategy Nash Equilibrium (PSNE)—an entry that is the smallest in its row and the largest in its column—it may not necessarily be a strict PSNE. However, if we are given a set of values $\mathcal{V}$ that includes the value of the game $V_{A}^{\star}$ , we can modify the previously constructed matrix to ensure that the PSNE becomes a strict PSNE. Our randomized algorithm for finding the strict PSNE has the additional advantage that, whenever the matrix contains a PSNE, it will, with high probability, compute a set that includes the value of the game $V_{A}^{\star}$ .
Now, we describe how to make the PSNE that maps to the optimal sub-matrix (which we will refer to as the optimal PSNE) a strict PSNE. Consider an entry that is in the column of the optimal PSNE and has a value equal to $V_{A}^{\star}$ . If this entry is not the optimal PSNE, using Lemma 1, we can show that the sub-matrix it maps to lacks a Nash equilibrium with all the rows in its support, although it does have a Nash equilibrium with all the columns in its support. Therefore, if we devise a method to decrease the values of all such entries by a very small amount $\varepsilon$ , the optimal PSNE will become strictly greater than all the entries in its column.
Although we do not know which column contains the optimal PSNE, we can decrease the values of entries by a very small amount $\varepsilon$ if the entry value is in the set $\mathcal{V}$ , which includes $V_{A}^{\star}$ , and the sub-matrix it maps to lacks a Nash equilibrium with all the rows in its support, but does have a Nash equilibrium with all the columns in its support. The values of entries corresponding to the optimal sub-matrix remain unchanged, as there exists a Nash equilibrium strategy with all rows and columns in its support. By modifying the entries in this way, we can ensure that the optimal PSNE is strictly larger than all other entries in its column. Similarly, we can make this PSNE strictly smaller than all other entries in its row. Note that while we modify other entries whose value is not equal to $V_{A}^{\star}$ , we ensure that the relative ordering with respect to $V_{A}^{\star}$ remains unchanged. Consequently, we implicitly construct a large matrix with a strict PSNE that maps to the optimal sub-matrix, and we can find this strict PSNE by querying nearly $n\cdot\text{poly}(k)$ entries of $A$ .
In summary, our main innovative technique is to reduce the problem of finding the set of all Nash equilibria to finding a strict PSNE in a large matrix, and then to locate this strict PSNE in a query-efficient manner.
We also provide a simple lower bound to demonstrate that $k=\max\{k_{1},k_{2}\}$ accurately captures the query complexity of the problem.
**Theorem 2**
*Let $\mathcal{A}$ be any randomized algorithm that correctly computes the set of all Nash equilibria $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . Let $\tau(M)$ denote the number of entries queried from the matrix $M$ by $\mathcal{A}$ . Then, there exists an input matrix $A\in\mathbb{R}^{n\times n}$ with $\max\{k_{1},k_{2}\}=k$ and $\min\{k_{1},k_{2}\}=1$ such that $\mathbb{E}[\tau(A)]\geq\frac{(n-1)(k+1)}{2}$ , where the randomness in $\tau(A)$ is due to $\mathcal{A}$ only.*
This result is not surprising, as changing the elements in the rows $\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ and columns $\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)$ can alter the set of all Nash equilibria, making it necessary to observe them.
### 1.3 Related work
Unlike the computational complexity literature, which focuses on minimizing the number of computational steps for algorithms that compute equilibrium while knowing the payoffs, the works on query complexity consider a scenario where an algorithm aims to compute an equilibrium for a game without knowing the payoffs beforehand. The main focus of such works is to minimize the number of payoff queries used. We discuss some of these works here.
Query complexity of zero-sum games: Recently, there has been renewed interest in the topic of query complexity of zero-sum games. Under the assumption that input matrix $A\in\{0,1\}^{n\times n}$ has a unique Nash equilibrium, Auger et al. [2] designed a randomized algorithm for finding the unique Nash equilibrium of $A$ by querying $O(n\log(n)\cdot k^{4k})$ entries of $A$ , where $k=|\operatorname*{supp}(x^{\star})|$ . Recall that for the special case when the input matrix has a strict PSNE, Bienstock et al. [5] presented a deterministic algorithm that finds the strict PSNE by querying $O(n)$ entries of the input matrix $A\in\mathbb{R}^{n\times n}$ . Recently, Dallant et al. [10] improved this result by designing a deterministic algorithm that finds the strict PSNE in $O(n\log^{*}n)$ time by querying $O(n)$ entries of the input matrix $A\in\mathbb{R}^{n\times n}$ . Parallel to our work, under the same special case, Dallant et al. [11] and Maiti et al. [21] designed randomized algorithms aimed at minimizing the time complexity and the statistical complexity (when the feedback on the payoff matrix is noisy) respectively.
Query complexity of general-sum games: The query complexity of finding various equilibrium concepts in games—including general-sum multi-player games—has been extensively studied; see the survey by Babichenko [4] for a comprehensive overview. For instance, Babichenko [3] demonstrated that the query complexity of computing an $\varepsilon$ -well-supported Nash equilibrium in an $n$ -player binary action game is $2^{\Omega(n)}$ .
Fearnley et al. [18] showed that in an $n\times n$ bimatrix game with payoffs in the range [0,1], a $\frac{1}{2}$ -approximate Nash equilibrium can be computed using at most $2n-1$ payoff queries. They complemented this result by showing that there is a constant $\varepsilon<\frac{1}{2}$ such that the query complexity of computing an $\varepsilon$ -approximate Nash equilibrium is more than $2n-1$ .
Goldberg and Roth [19] proved that for a game with $n$ players and $m$ actions, the query complexity of finding an $\varepsilon$ -approximate coarse correlated equilibrium is $O\left(\frac{nm\log m}{\varepsilon^{2}}\right)$ . A randomized algorithm was designed by Fearnley and Savani [17] that finds a $\left(\frac{3-\sqrt{5}}{2}+\varepsilon\right)$ -approximate Nash equilibrium in $n\times n$ bimatrix games using $O\left(\frac{n\log n}{\varepsilon^{2}}\right)$ payoff queries. Later, Deligkas et al. [16] presented an algorithm that finds a $\frac{1}{2}+\varepsilon$ -well-supported Nash equilibrium in $n\times n$ general-sum matrix games using $O\left(\frac{n\log n}{\varepsilon^{4}}\right)$ payoff queries.
### 1.4 Notation
We refer to the entry in the $i$ th row and $j$ th column of a matrix $M$ as entry $(i,j)$ of $M$ , and we denote its value by $M_{i,j}$ . Let $e_{i}$ denote the standard basis vector with the $i$ th entry equal to one and all other entries equal to zero. We use $\mathbf{1}$ to represent the all-ones vector.
Let $[n]:=\{1,2,\ldots,n\}$ , and let $\binom{[n]}{k_{0}}$ denote the set of all subsets of $[n]$ of size $k_{0}$ . For any $v\in\mathbb{R}^{n}$ , let $\operatorname*{supp}(v):=\{i\in[n]\mid v_{i}\neq 0\}$ denote the support of $v$ .
For any matrix $M\in\mathbb{R}^{n\times m}$ , let $V_{M}^{\star}:=\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt ,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height =3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim= 0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.219 4pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[hei ght=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{m}}\langle x,My\rangle$ denote the value of the zero-sum game on matrix $M$ . We say an entry $(i_{\star},j_{\star})$ is a Pure Strategy Nash equilibrium (in short PSNE) of $M$ if $M_{i,j_{\star}}\leq M_{i_{\star},j_{\star}}\leq M_{i_{\star},j}$ for all $i\neq i_{\star},j\neq j_{\star}$ . We say an entry $(i_{\star},j_{\star})$ is a strict PSNE of $M$ if $M_{i,j_{\star}}<M_{i_{\star},j_{\star}}<M_{i_{\star},j}$ for all $i\neq i_{\star},j\neq j_{\star}$ .
Let $\mathbb{R}^{\mathcal{I}}$ represent the space of all real-valued vectors indexed by the set of indices $\mathcal{I}$ . The simplex over a set of indices $\mathcal{I}$ , denoted $\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{\mathcal{I}}$ , is defined as the set of vectors $x\in\mathbb{R}^{\mathcal{I}}$ such that $\sum_{i\in\mathcal{I}}x_{i}=1$ and $x_{i}\geq 0$ for all $i\in\mathcal{I}$ .
For simplicity, we use $\log x$ to denote $\log_{2}x$ . Also for simplicity, unless otherwise noted, by number of entries queried, we mean the number of distinct entries queried.
Outline of the paper:
In Section 2, we discuss key properties of Nash equilibria. In Section 3, we present important technical lemmas that will be utilized throughout the paper. Section 4 introduces the randomized algorithm for finding strict PSNE, while Section 5 details the algorithm for finding the set of all Nash equilibria and provides a lower bound on query complexity.
## 2 Properties of Nash Equilibrium
We now state some properties of Nash equilibrium (see [20] and [6] for more details). Consider $A\in\mathbb{R}^{n\times n}$ . The pair $(x_{\star},y_{\star})$ is a Nash equilibrium of $A$ if and only if the following hold simultaneously:
- The value of the game on $A$ satisfies $\langle e_{i},Ay_{\star}\rangle=V^{\star}_{A}$ for any $i\in\operatorname*{supp}(x_{\star})$ and $\langle x_{\star},Ae_{j}\rangle=V^{\star}_{A}$ for any $j\in\operatorname*{supp}(y_{\star})$ .
- The value of the game on $A$ satisfies $\langle e_{i},Ay_{\star}\rangle\leq V^{\star}_{A}$ for any $i\notin\operatorname*{supp}(x_{\star})$ and $\langle x_{\star},Ae_{j}\rangle\geq V^{\star}_{A}$ for any $j\notin\operatorname*{supp}(y_{\star})$ .
It is also easy to observe that $(x_{\star},y_{\star})$ is a Nash equilibrium of $A$ if and only if $\langle x_{\star},Ay_{\star}\rangle=\max_{i\in[n]}\langle e_{i},Ay_{\star} \rangle=\min_{j\in[n]}\langle x_{\star},Ae_{j}\rangle$ . Now, recall that due to Von Neumann’s minimax theorem, we have
$$
V_{A}^{\star}:=\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0
pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt
,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height
=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=
0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.219
4pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[hei
ght=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}_{n}}\langle x,Ay\rangle=\min_{y\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle.
$$
Observe that this is also equivalent to
$$
V_{A}^{\star}=\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0
pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt
,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height
=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}_{n}}\min_{j\in[n]}\langle x,Ae_{j}\rangle=\min_{y\in\mathchoice{
\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{
graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0
pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt
30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{i\in[n]}\langle e_{i},Ay\rangle.
$$
Also recall that $(x_{\star},y_{\star})$ is a Nash equilibrium of the matrix $A$ if and only if
$$
x_{\star}\in\mathcal{X}_{\star}:=\arg\max_{x\in\mathchoice{\includegraphics[he
ight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle\quad\text{and}\quad y
_{\star}\in\mathcal{Y}_{\star}:=\arg\min_{y\in\mathchoice{\includegraphics[hei
ght=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[h
eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle.
$$
This implies that if there are strategies $x_{\star},y_{\star}\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt, trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height= 4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}$ such that $\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41 112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x_{ \star},Ay\rangle=V_{A}^{\star}$ and $\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41 112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay_{ \star}\rangle=V_{A}^{\star}$ , then $x_{\star}\in\mathcal{X}_{\star}$ and $y_{\star}\in\mathcal{Y}_{\star}$ . Also, convex combination of multiple Nash equilibria is a Nash equilibrium as $\mathcal{X}_{\star},\mathcal{Y}_{\star}$ are compact convex sets.
The sets $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ can be efficiently represented as the convex hulls of their extreme points, which are finite and non-zero in number. Let $\mathcal{E}(\mathcal{X})$ denote the set of extreme points of a set $\mathcal{X}$ . For any submatrix $\tilde{A}$ of $A$ , define $\tilde{x}$ and $\tilde{y}$ as the vectors formed by removing the components of $x$ and $y$ that correspond to the rows and columns missing from $\tilde{A}$ . The following lemma describes an important property of the extreme points of $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ .
**Lemma 2 (Bohnenblust et al. [6])**
*A Nash equilibrium $(x_{\star},y_{\star})$ of $A$ is in the set $\mathcal{E}(\mathcal{X}_{\star})\times\mathcal{E}(\mathcal{Y}_{\star})$ if and only if there is a square sub-matrix $\tilde{A}$ of $A$ such that
$$
\tilde{x}^{\top}_{\star}=\frac{\mathbf{1}^{\top}adj(\tilde{A})}{\mathbf{1}^{
\top}adj(\tilde{A})\mathbf{1}},\;\tilde{y}_{\star}=\frac{adj(\tilde{A})\mathbf
{1}}{\mathbf{1}^{\top}adj(\tilde{A})\mathbf{1}},\;V_{A}^{\star}=\frac{det(
\tilde{A})}{\mathbf{1}^{\top}adj(\tilde{A})\mathbf{1}}
$$*
It is easy to observe that we can compute all the extreme points of $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ by examining all the square submatrices $\tilde{A}$ of the input matrix $A$ and applying the above lemma. Let $I_{x}=\{j\in[n]:V_{A}^{\star}=\langle x,Ae_{j}\rangle\}$ and $I_{y}=\{i\in[n]:V_{A}^{\star}=\langle e_{i},Ay\rangle\}$ . Since the sets $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ are the convex hulls of $\mathcal{E}(\mathcal{X}_{\star})$ and $\mathcal{E}(\mathcal{Y}_{\star})$ , we have $\bigcup\limits_{x_{\star}\in\mathcal{X}_{\star}}\operatorname*{supp}(x_{\star} )=\bigcup\limits_{x_{\star}\in\mathcal{E}(\mathcal{X}_{\star})}\operatorname*{ supp}(x_{\star})$ , $\bigcup\limits_{y_{\star}\in\mathcal{Y}_{\star}}\operatorname*{supp}(y_{\star} )=\bigcup\limits_{y_{\star}\in\mathcal{E}(\mathcal{Y}_{\star})}\operatorname*{ supp}(y_{\star})$ , $\bigcap\limits_{x_{\star}\in\mathcal{X}_{\star}}I_{x}=\bigcap\limits_{x_{\star }\in\mathcal{E}(\mathcal{X}_{\star})}I_{x}$ , and $\bigcap\limits_{y_{\star}\in\mathcal{Y}_{\star}}I_{y}=\bigcap\limits_{y_{\star }\in\mathcal{E}(\mathcal{Y}_{\star})}I_{y}$ . This fact allows us to verify conditions involving $I_{x}$ , $I_{y}$ , and the union of supports, which is critical to our algorithm in later sections.
## 3 Technical Lemmas for Nash Equilibrium
Recall that for the input matrix $A\in\mathbb{R}^{n\times n}$ , $\mathcal{X}_{\star}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ and $\mathcal{Y}_{\star}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ . Also recall that $\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{\mathcal{I}}$ represents the simplex over a set of indices $\mathcal{I}$ . For any vector $v$ and the set of indices $\mathcal{S}=\{i_{1},\ldots,i_{\ell}\}$ , let $v_{\mathcal{S}}$ denote the vector $(v_{i_{1}},\ldots,v_{i_{\ell}})$ . We now state an important lemma that provides the necessary and sufficient condition for a subset of rows and columns to be optimal.
**Lemma 3**
*Let $\mathcal{I}$ and $\mathcal{J}$ be two subsets of $[n]$ . Let $M_{1}$ be the submatrix of $A$ with row indices $\mathcal{I}$ and column indices $[n]$ , and $M_{2}$ be the submatrix of $A$ with row indices $[n]$ and column indices $\mathcal{J}$ . Let $\mathcal{X}_{M_{1}}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{\mathcal{I}}}\min_{y\in\mathchoice{\includegraphics[h eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,M_{1}y\rangle$ and $\mathcal{Y}_{M_{2}}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{\mathcal{J}}}\max_{x\in\mathchoice{\includegraphics[h eight=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,M_{2}y\rangle$ . Let $\mathcal{X}_{A}:=\{x\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}:x_{\mathcal{I}}\in\mathcal{X}_{M_{1}}\}$ and $\mathcal{Y}_{A}:=\{y\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}:y_{\mathcal{J}}\in\mathcal{Y}_{M_{2}}\}$ . The following holds:
1. If $\mathcal{I}=\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ and $\mathcal{J}=\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)$ , then the following holds:
- $\mathcal{X}_{A}=\mathcal{X}_{\star}$ and $\mathcal{Y}_{A}=\mathcal{Y}_{\star}$
- $V_{M_{1}}^{\star}=V_{A}^{\star}$ and $\bigcap\limits_{x\in\mathcal{X}_{A}}I_{x}=\mathcal{J}$ where $I_{x}=\{j\in[n]:V_{A}^{\star}=\langle x,Ae_{j}\rangle\}$ .
- $V_{M_{2}}^{\star}=V_{A}^{\star}$ and $\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}=\mathcal{I}$ where $I_{y}=\{i\in[n]:V_{A}^{\star}=\langle e_{i},Ay\rangle\}$ .
1. Let $I_{x}^{\prime}=\{j\in[n]:V_{M_{1}}^{\star}=\langle x,Ae_{j}\rangle\}$ and $I_{y}^{\prime}=\{i\in[n]:V_{M_{2}}^{\star}=\langle e_{i},Ay\rangle\}$ . If $V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ , $\bigcap\limits_{x\in\mathcal{X}_{A}}I_{x}^{\prime}=\mathcal{J}$ and $\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}^{\prime}=\mathcal{I}$ then $\mathcal{X}_{A}=\mathcal{X}_{\star}$ and $\mathcal{Y}_{A}=\mathcal{Y}_{\star}$ .*
* Proof*
Let us assume that $\mathcal{I}=\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)$ and $\mathcal{J}=\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)$ . First consider $\hat{x}\in\mathcal{X}_{A}$ . Observe that $\operatorname*{supp}(\hat{x})\subseteq\mathcal{I}$ . Now observe that $V_{M_{1}}^{\star}=\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim= 0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.219 4pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[hei ght=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}}\langle\hat{x},Ay\rangle\leq\max_{x\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle=V_{A}^{\star}$ . Next consider $x^{\star}\in\mathcal{X}_{\star}$ . Observe that $\operatorname*{supp}(x^{\star})\subseteq\mathcal{I}$ . Now observe that $V_{A}^{\star}=\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt ,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height =3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}}\langle x^{\star}_{\mathcal{I}},M_{1}y\rangle\leq\max_{x\in \mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{\mathcal{I}}}\min_{y\in \mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,M_{1}y \rangle=V_{M_{1}}^{\star}$ . Hence, we have $V_{A}^{\star}=V_{M_{1}}^{\star}$ and therefore $\hat{x}\in\mathcal{X}_{\star}$ and $x^{\star}\in\mathcal{X}_{A}$ implying that $\mathcal{X}_{A}=\mathcal{X}_{\star}$ . Analogously, one can show that $V_{M_{2}}^{\star}=V_{A}^{\star}$ and $\mathcal{Y}_{A}=\mathcal{Y}_{\star}$ . Due to Lemma 1, we get that $\bigcap\limits_{x\in\mathcal{X}_{A}}I_{x}=\bigcap\limits_{x\in\mathcal{X}_{ \star}}I_{x}=\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)= \mathcal{J}$ and $\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}=\bigcap\limits_{y\in\mathcal{Y}_{ \star}}I_{y}=\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)= \mathcal{I}$ . Next let us assume that $V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ , $\bigcap\limits_{x\in\mathcal{X}_{A}}I_{x}^{\prime}=\mathcal{J}$ and $\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}^{\prime}=\mathcal{I}$ . Consider $\hat{x}\in\mathcal{X}_{A}$ and $\hat{y}\in\mathcal{Y}_{A}$ . Observe that $\operatorname*{supp}(\hat{x})\subseteq\mathcal{I}$ and $\operatorname*{supp}(\hat{y})\subseteq\mathcal{J}$ . Now observe that $\langle\hat{x},A\hat{y}\rangle\geq\min_{y\in\mathchoice{\includegraphics[heigh t=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle\hat{x},Ay\rangle=V_{M_{1}}^{\star}$ . Next observe that $\langle\hat{x},A\hat{y}\rangle\leq\max_{x\in\mathchoice{\includegraphics[heigh t=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,A\hat{y}\rangle=V_{M_{2}}^{\star}$ . Since $V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ , we have $\langle\hat{x},A\hat{y}\rangle=V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ . This implies that $\langle\hat{x},A\hat{y}\rangle=V_{M_{2}}^{\star}=\max_{x\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,A\hat{y}\rangle$ and $\langle\hat{x},A\hat{y}\rangle=V_{M_{1}}^{\star}=\min_{y\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle\hat{x},Ay\rangle$ . Hence $(\hat{x},\hat{y})$ is a Nash equilibrium of $A$ and $V_{A}^{\star}=\langle\hat{x},A\hat{y}\rangle=V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ . This implies that $\mathcal{X}_{A}\subseteq\mathcal{X}_{\star}$ and $\mathcal{Y}_{A}\subseteq\mathcal{Y}_{\star}$ . Consider $x^{\star}\in\mathcal{X}_{\star}$ . Due to Lemma 1, we have $\operatorname*{supp}(x^{\star})\subseteq\bigcap\limits_{y\in\mathcal{Y}_{\star }}I_{y}\subseteq\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}^{\prime}=\mathcal{I}$ . Now we have $\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25. 0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim =0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41 112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x^{ \star}_{\mathcal{I}},M_{1}y\rangle=\min_{y\in\mathchoice{\includegraphics[heig ht=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0. 0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x^{\star},Ay\rangle=V_{A}^{\star }=V_{M_{1}}^{\star}$ . Hence, $x^{\star}_{\mathcal{I}}\in\mathcal{X}_{M_{1}}$ which implies $\mathcal{X}_{\star}\subseteq\mathcal{X}_{A}$ . Analogously, one can show that $\mathcal{Y}_{\star}\subseteq\mathcal{Y}_{A}$ . Hence, we have $\mathcal{X}_{A}=\mathcal{X}_{\star}$ and $\mathcal{Y}_{A}=\mathcal{Y}_{\star}$ . ∎
The following lemma states that a matrix can have at most one strict PSNE.
**Lemma 4**
*Any matrix $M\in\mathbb{R}^{n\times m}$ can have at most one strict PSNE.*
* Proof*
Consider a matrix $M$ with a strict PSNE at $(i_{\star},j_{\star})$ . Since $(i_{\star},j_{\star})$ is a strict PSNE, we have $M_{i,j_{\star}}<M_{i_{\star},j}$ for all $(i,j)\neq(i_{\star},j_{\star})$ . Suppose there is another entry $(i^{\prime},j^{\prime})\neq(i_{\star},j_{\star})$ that is also a strict PSNE. Since $(i^{\prime},j^{\prime})$ is a PSNE, we would have $M_{i_{\star},j^{\prime}}\leq M_{i^{\prime},j^{\prime}}\leq M_{i^{\prime},j_{ \star}}$ . However, this contradicts the inequality $M_{i^{\prime},j_{\star}}<M_{i_{\star},j^{\prime}}$ . Therefore, the matrix $M$ cannot have more than one strict PSNE. ∎
## 4 An important subroutine for PSNE
A key part of our algorithm for finding the set of all Nash equilibria is a query-efficient subroutine for a matrix $M\in\mathbb{R}^{n\times m}$ that has a PSNE and potentially a strict PSNE. Moreover, we assume that we have access to the matrix $M$ through a query oracle $\mathcal{Q}$ which operates as follows:
- A single call to query oracle $\mathcal{Q}$ with input $(\text{row},i)$ returns the entire $i$ -th row of $M$ .
- A single call to query oracle $\mathcal{Q}$ with input $(\text{column},j)$ returns the entire $j$ -th column of $M$ .
In this section, we work with a matrix $M\in\mathbb{R}^{n\times m}$ and we aim to achieve the following goals by making $O(\log^{2}(n+m))$ calls to the query oracle $\mathcal{Q}$ :
- If the matrix $M$ has a PSNE, we construct a set of distinct real numbers that contains $V_{M}^{\star}$ with high probability.
- If the matrix $M$ has a strict PSNE, we find the strict PSNE with high probability.
### 4.1 A Randomized Algorithm to Find the Strict PSNE
In this section, we assume that the input matrix $M\in\mathbb{R}^{n\times m}$ has a strict PSNE $(i_{\star},j_{\star})$ . By definition, we have that $M_{i,j_{\star}}<M_{i_{\star},j_{\star}}<M_{i_{\star},j}$ for all $i\in[n]\setminus\{i_{\star}\}$ and $j\in[m]\setminus\{j_{\star}\}$ . We use this crucial fact to design a randomized algorithm, FindPsne (Algorithm 1), for finding the strict PSNE.
We first describe our procedure, FindPsne, intuitively. The procedure progresses through a logarithmic number of stages. In each stage, it aims to eliminate half of the sub-optimal rows and columns. At any given stage where the goal is to eliminate half of the sub-optimal rows, we begin by sampling a logarithmic number of row indices and querying all entries in each sampled row. Next, we examine the highest values among the queried entries in each column and select the column $\widehat{j}$ with the lowest of these values. We then sample all entries in column $\widehat{j}$ and remove rows where the corresponding entries in the column $\widehat{j}$ are below the median value of that column. An analogous process exists to eliminate half of the sub-optimal columns. This process of sampling, querying, and reducing the search space continues until only one row and one column remain, which together form the strict PSNE. For a formal description of the procedure FindPsne, please refer to Algorithm 1.
1 $n\leftarrow$ number of rows of $M$ , $m\leftarrow$ number of columns of $M$ , $\mathcal{X}_{1}\leftarrow[n],\>\mathcal{Y}_{1}\leftarrow[m],\>\mathcal{V}_{1}\leftarrow\emptyset$
2 for $t=1,2,\ldots$ do
3 if $|\mathcal{X}_{t}|=1$ and $|\mathcal{Y}_{t}|=1$ then
4 $\mathcal{V}\leftarrow\mathcal{V}_{t}\cup\{M_{\widehat{i},\widehat{j}}\}$ , where $\mathcal{X}_{t}=\{\;\widehat{i}\;\}$ and $\mathcal{Y}_{t}=\{\;\widehat{j}\;\}$
5 return $(\;\widehat{i},\>\widehat{j}\;)$ and $\mathcal{V}$
6 if $|\mathcal{X}_{t}|=1$ then
7 Query all the elements of row $\widehat{i}$ , where $\mathcal{X}_{t}=\{\;\widehat{i}\;\}$
8 Choose the argument $\widehat{j}\in\operatorname*{arg\,min}_{j\in\mathcal{Y}_{t}}M_{\widehat{i},j}$ and set $\mathcal{V}\leftarrow\mathcal{V}_{t}\cup\{M_{\widehat{i},\widehat{j}}\}$
9 return $(\;\widehat{i},\>\widehat{j}\;)$ and $\mathcal{V}$
10 if $|\mathcal{Y}_{t}|=1$ then
11 Query all the elements of column $\widehat{j}$ , where $\mathcal{Y}_{t}=\{\;\widehat{j}\;\}$
12 Choose the argument $\widehat{i}\in\operatorname*{arg\,max}_{i\in\mathcal{X}_{t}}M_{i,\widehat{j}}$ and set $\mathcal{V}\leftarrow\mathcal{V}_{t}\cup\{M_{\widehat{i},\widehat{j}}\}$
13 return $(\;\widehat{i},\>\widehat{j}\;)$ and $\mathcal{V}$
14 $n_{t}\leftarrow|\mathcal{X}_{t}|$ , $m_{t}\leftarrow|\mathcal{Y}_{t}|$ and $\ell\leftarrow\left\lceil 2\log\left(\frac{2(n+m)^{2}}{\delta}\right)\right\rceil$
15 $\overline{X}_{t}\leftarrow\{x_{1},x_{2},\ldots,x_{\ell}\}$ ; $x_{i}$ ’s are drawn independently and uniformly at random from $\mathcal{X}_{t}$
16 $\overline{Y}_{t}\leftarrow\{y_{1},y_{2},\ldots,y_{\ell}\}$ ; $y_{j}$ ’s are drawn independently and uniformly at random from $\mathcal{Y}_{t}$
17 Query all the entries of $M$ in $\{(i,j):i\in\mathcal{X}_{t},j\in\overline{Y}_{t}\}\cup\{(i,j):i\in\overline{X} _{t},j\in\mathcal{Y}_{t}\}$
18
19 Choose the arguments $\widehat{i}\in\operatorname*{arg\,max}_{i\in\mathcal{X}_{t}}\min_{j\in \overline{Y}_{t}}M_{i,j}$ and $\widehat{j}\in\operatorname*{arg\,min}_{j\in\mathcal{Y}_{t}}\max_{i\in \overline{X}_{t}}M_{i,j}$
20 $\mathcal{V}_{t+1}\leftarrow\mathcal{V}_{t}\cup\{\max_{i\in\mathcal{X}_{t}}\min _{j\in\overline{Y}_{t}}M_{i,j},\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X }_{t}}M_{i,j}\}$
21 Query all the entries of $M$ in $\{({i,\widehat{j}}):i\in\mathcal{X}_{t}\}\cup\{({\widehat{i},\>j}):j\in \mathcal{Y}_{t}\}$
22
23 $\mathcal{X}_{t+1}\leftarrow\{i_{1},i_{2},\ldots,i_{\lfloor n_{t}/2\rfloor}\} \subseteq\mathcal{X}_{t}=\{i_{1},\ldots,i_{n_{t}}\}$ , where $M_{i_{1},\widehat{j}}\geq M_{i_{2},\widehat{j}}\geq\ldots\geq M_{i_{n_{t}}, \widehat{j}}$
24 $\mathcal{Y}_{t+1}\leftarrow\{j_{1},j_{2},\ldots,j_{\lfloor m_{t}/2\rfloor}\} \subseteq\mathcal{Y}_{t}=\{j_{1},\ldots,j_{m_{t}}\}$ , where $M_{\widehat{i},j_{1}}\leq M_{\widehat{i},j_{2}}\leq\ldots\leq M_{\widehat{i},j _{m_{t}}}$
25
Algorithm 1 FindPsne ( $M,\delta$ )
We now begin the analysis of Algorithm 1 with the following proposition.
**Proposition 1**
*Fix $c,k\in\mathbb{N}$ with $k\geq 2$ . Consider a set $S=\{a_{1},a_{2},\ldots,a_{k}\}$ , and let $S_{1}=\{a_{1},a_{2},\ldots,a_{\lfloor k/2\rfloor}\}$ . Let $\ell=\lceil 2\log(\frac{c}{\delta})\rceil$ , and let $X=\{x_{1},x_{2},\ldots,x_{\ell}\}$ where $x_{i}$ ’s are drawn independently and uniformly at random from $S$ . Then, $X\cap S_{1}\neq\emptyset$ with probability at least $1-\frac{\delta}{c}$ .*
* Proof*
As the events $\{x_{i}\notin S_{1}\}$ are independent,
$$
\displaystyle\mathbb{P}(X\cap S_{1}=\emptyset)=\prod_{i=1}^{\ell}\mathbb{P}(x_
{i}\notin S_{1})=\left\lparen\tfrac{k-\lfloor k/2\rfloor}{k}\right\rparen^{
\ell}\leq\left\lparen\tfrac{2}{3}\right\rparen^{2\log(\frac{c}{\delta})}<\left
\lparen\tfrac{1}{2}\right\rparen^{\log(\frac{c}{\delta})}=\tfrac{\delta}{c}. \tag{23}
$$
∎
We now establish the query complexity of FindPsne in the following lemma.
**Lemma 5**
*FindPsne $(A,\delta)$ can be executed by calling the query oracle $\mathcal{Q}$ at most $O(\log(n+m)\cdot\log(\frac{n+m}{\delta}))$ times.*
* Proof*
Observe that the algorithm terminates at a fixed iteration $t_{\star}$ which is at most $\min\{\log(n),\log(m)\}+1$ . In each iteration $t<t_{\star}$ , line 1 requires at most $2\ell$ calls to the query oracle $\mathcal{Q}$ and line 1 requires two calls to the query oracle $\mathcal{Q}$ . Final iteration $t_{\star}$ requires at most one call to the query oracle $\mathcal{Q}$ . Hence, FindPsne $(A,\delta)$ can be executed by calling the query oracle $\mathcal{Q}$ at most $O(\log(n+m)\cdot\log(\frac{n+m}{\delta}))$ times. ∎
Finally, we establish the correctness of FindPsne in the following lemma.
**Lemma 6**
*If $M$ has a strict PSNE, then FindPsne $(M,\delta)$ returns the strict PSNE with probability at least $1-\delta$ .*
* Proof*
Let $(i_{\star},j_{\star})$ be the strict PSNE of $M$ . Consider an iteration $t$ such that $\min\{|\mathcal{X}_{t}|,|\mathcal{Y}_{t}|\}\geq 2$ . Let us assume that $i_{\star}\in\mathcal{X}_{t}$ and $j_{\star}\in\mathcal{Y}_{t}$ . Under this assumption, we will show that $i_{\star}\in\mathcal{X}_{t+1}$ and $j_{\star}\in\mathcal{Y}_{t+1}$ with probability at least $1-\frac{\delta}{n+m}$ . For each $j\in\mathcal{Y}_{t}$ , let us relabel the indices in $\mathcal{X}_{t}$ as $\{i_{1},i_{2},\ldots,i_{n_{t}}\}$ such that $M_{i_{1},j}\geq M_{i_{2},j}\geq\ldots\geq M_{i_{n_{t}},j}$ and define $\mathcal{R}_{t,j}:=\{i_{1},i_{2},\ldots,i_{\lfloor n_{t}/2\rfloor}\}$ . Recall that $\overline{X}_{t}$ is a random subset of $\mathcal{X}_{t}$ . By Proposition 1 we have $\mathbb{P}(\overline{X}_{t}\cap\mathcal{R}_{t,j}\neq\emptyset)\geq 1-\frac{ \delta}{2(n+m)^{2}}$ for a fixed $j$ . By the union bound, we then have that with probability at least $1-\frac{\delta}{2(n+m)}$ , the event $\mathcal{E}_{t}:=\{\overline{X}_{t}\cap\mathcal{R}_{t,j}\neq\emptyset\>\forall \>j\in\mathcal{Y}_{t}\}$ holds. Note that if $\mathcal{E}_{t}$ holds, for each $j\in\mathcal{Y}_{t}$ we have $\operatorname*{arg\,max}_{i\in\overline{X}_{t}}M_{i,j}\cap\mathcal{R}_{t,j}\neq\emptyset$ . Assume $\mathcal{E}_{t}$ holds, and let $v^{\star}_{y}=\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}$ . Recall that $\widehat{j}\in\arg\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}$ . If $\widehat{j}=j_{\star}$ , then $i_{\star}\in\mathcal{R}_{t,\widehat{j}}=\mathcal{X}_{t+1}$ , as $M_{i_{\star},j_{\star}}$ is the unique highest element of column $j_{\star}$ . Next, let us consider the case where $\widehat{j}\neq j_{\star}$ . As $\mathcal{E}_{t}$ holds, $\operatorname*{arg\,max}_{i\in\overline{X}_{t}}M_{i,\widehat{j}}\cap\mathcal{R }_{t,\widehat{j}}\neq\emptyset$ . We also see that $v^{\star}_{y}=\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}\leq \max_{i\in\overline{X}_{t}}M_{i,j_{\star}}\leq M_{i_{\star},j_{\star}}$ . Recall that for all $j\in\mathcal{Y}_{t}\setminus\{j_{\star}\}$ , we have $M_{i_{\star},j}>M_{i_{\star},j_{\star}}$ . Now for all $i\in\mathcal{X}_{t}\setminus\mathcal{R}_{t,\widehat{j}}$ , we have $M_{i,\widehat{j}}\leq v_{y}^{\star}\leq M_{i_{\star},j_{\star}}<M_{i_{\star}, \widehat{j}}$ . Hence, $i_{\star}\in\mathcal{R}_{t,\widehat{j}}=\mathcal{X}_{t+1}$ . By an identical argument, we have $j_{\star}\in\mathcal{Y}_{t+1}$ with probability at least $1-\frac{\delta}{2(n+m)}$ . Hence, by the union bound, with probability at least $1-\frac{\delta}{n+m}$ we have $i_{\star}\in\mathcal{X}_{t+1}$ and $j_{\star}\in\mathcal{Y}_{t+1}$ . Observe that the algorithm terminates at a fixed iteration $t_{\star}$ which is at most $\min\{\log(n),\log(m)\}+1$ . Now observe that FindPsne $(M,\delta)$ returns $(i_{\star},j_{\star})$ if $i_{\star}\in\mathcal{X}_{t_{\star}}$ and $j_{\star}\in\mathcal{Y}_{t_{\star}}$ as by definition the entry $(i_{\star},j_{\star})$ is the unique least element in the row $i_{\star}$ and unique highest element in the column $j_{\star}$ . Let $p_{t}=\mathbb{P}(i_{\star}\in\mathcal{X}_{t},j_{\star}\in\mathcal{Y}_{t}|i_{ \star}\in\mathcal{X}_{t-1},j_{\star}\in\mathcal{Y}_{t-1})$ . Now we have the following due to the chain rule and Bernoulli’s inequality,
$$
\mathbb{P}(i_{\star}\in\mathcal{X}_{t_{\star}},j_{\star}\in\mathcal{Y}_{t_{
\star}})=\prod_{t=2}^{t_{\star}}p_{t}\geq\left(1-\frac{\delta}{n+m}\right)^{t_
{\star}}\geq 1-\delta.
$$
∎
### 4.2 Important property of FindPsne for Non-Strict PSNE
Let us assume we are given an input matrix $M$ with a PSNE $(i_{\star},j_{\star})$ . By definition, we have $M_{i,j_{\star}}\leq M_{i_{\star},j_{\star}}\leq M_{i_{\star},j}$ for all $i\in[n]\setminus\{i_{\star}\}$ and $j\in[m]\setminus\{j_{\star}\}$ . In the following lemma, we state an important property of FindPsne, applicable even when the PSNE is not strict.
**Lemma 7**
*Consider an input matrix $M$ which has a PSNE $(i_{\star},j_{\star})$ . With probability at least $1-\delta$ , $\textsc{FindPsne}(M,\delta)$ returns a set $\mathcal{V}$ of distinct real numbers such that $V_{M}^{\star}\in\mathcal{V}$ .*
* Proof*
Let $(i_{\star},j_{\star})$ be a PSNE of $M$ . Observe that the algorithm terminates at a fixed iteration $t_{\star}$ which is at most $\min\{\log(n),\log(m)\}+1$ . Consider an iteration $t<t_{\star}$ . Let us assume that $i_{\star}\in\mathcal{X}_{t}$ and $j_{\star}\in\mathcal{Y}_{t}$ . Under this assumption, we will show that with probability at least $1-\frac{\delta}{n+m}$ either $i_{\star}\in\mathcal{X}_{t+1}$ and $j_{\star}\in\mathcal{Y}_{t+1}$ or $V_{M}^{\star}\in\mathcal{V}_{t+1}$ . For each $j\in\mathcal{Y}_{t}$ , let us relabel the indices in $\mathcal{X}_{t}$ as $\{i_{1},i_{2},\ldots,i_{n_{t}}\}$ such that $M_{i_{1},j}\geq M_{i_{2},j}\geq\ldots\geq M_{i_{n_{t}},j}$ and define $\mathcal{R}_{t,j}:=\{i_{1},i_{2},\ldots,i_{\lfloor n_{t}/2\rfloor}\}$ . Recall that $\overline{X}_{t}$ is a random subset of $\mathcal{X}_{t}$ . By Proposition 1 we have $\mathbb{P}(\overline{X}_{t}\cap\mathcal{R}_{t,j}\neq\emptyset)\geq 1-\frac{ \delta}{2(n+m)^{2}}$ for a fixed $j$ . By the union bound, we then have that with probability at least $1-\frac{\delta}{2(n+m)}$ , the event $\mathcal{E}_{t}:=\{\overline{X}_{t}\cap\mathcal{R}_{t,j}\neq\emptyset\>\forall \>j\in\mathcal{Y}_{t}\}$ holds. Note that if $\mathcal{E}_{t}$ holds, for each $j\in\mathcal{Y}_{t}$ we have $\operatorname*{arg\,max}_{i\in\overline{X}_{t}}M_{i,j}\cap\mathcal{R}_{t,j}\neq\emptyset$ . Assume $\mathcal{E}_{t}$ holds, and let $v^{\star}_{y}=\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}$ . Recall that $\widehat{j}\in\arg\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}$ . As $\mathcal{E}_{t}$ holds, $\operatorname*{arg\,max}_{i\in\overline{X}_{t}}M_{i,\widehat{j}}\cap\mathcal{R }_{t,\widehat{j}}\neq\emptyset$ . We also see that $v^{\star}_{y}=\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}\leq \max_{i\in\overline{X}_{t}}M_{i,j_{\star}}\leq M_{i_{\star},j_{\star}}=V_{M}^{\star}$ . If $v^{\star}_{y}=V_{M}^{\star}$ , then $V_{M}^{\star}\in\mathcal{V}_{t+1}$ as $\min_{j\in\mathcal{Y}_{t}}\max_{i\in\overline{X}_{t}}M_{i,j}$ is added to $\mathcal{V}_{t}$ to create $\mathcal{V}_{t+1}$ . Recall that for all $j\in\mathcal{Y}_{t}$ , we have $M_{i_{\star},j}\geq M_{i_{\star},j_{\star}}$ . Now if $v^{\star}_{y}<V_{M}^{\star}$ , then for all $i\in\mathcal{X}_{t}\setminus\mathcal{R}_{t,\widehat{j}}$ , we have $M_{i,\widehat{j}}\leq v_{y}^{\star}<M_{i_{\star},j_{\star}}\leq M_{i_{\star}, \widehat{j}}$ . Hence, in this case we have $i_{\star}\in\mathcal{R}_{t,\widehat{j}}=\mathcal{X}_{t+1}$ . By an identical argument, with probability at least $1-\frac{\delta}{2(n+m)}$ , either $j_{\star}\in\mathcal{Y}_{t+1}$ or $V_{M}^{\star}\in\mathcal{V}_{t+1}$ . Hence, by the union bound, with probability at least $1-\frac{\delta}{n+m}$ , either $i_{\star}\in\mathcal{X}_{t+1}$ and $j_{\star}\in\mathcal{Y}_{t+1}$ or $V_{M}^{\star}\in\mathcal{V}_{t+1}$ . For the last iteration $t_{\star}$ , if $i_{\star}\in\mathcal{X}_{t_{\star}}$ and $j_{\star}\in\mathcal{Y}_{t_{\star}}$ , then $V_{M}^{\star}\in\mathcal{V}$ as the entry $(i_{\star},j_{\star})$ is the highest element of column $j_{\star}$ and the least element of row $i_{\star}$ . Let $p_{t}=\mathbb{P}(i_{\star}\in\mathcal{X}_{t}\text{ and }j_{\star}\in\mathcal{Y }_{t}\text{ or }V_{M}^{\star}\in\mathcal{V}_{t}|i_{\star}\in\mathcal{X}_{t-1} \text{ and }j_{\star}\in\mathcal{Y}_{t-1}\text{ or }V_{M}^{\star}\in\mathcal{V }_{t-1})$ . Now we have the following due to the chain rule and Bernoulli’s inequality,
$$
\mathbb{P}(V_{M}^{\star}\in\mathcal{V})=\prod_{t=2}^{t_{\star}}p_{t}\geq\left(
1-\frac{\delta}{n+m}\right)^{t_{\star}}\geq 1-\delta.
$$
∎
## 5 Results for finding the set of all Nash equilibria
Consider an input matrix $A\in\mathbb{R}^{n\times n}$ . Recall that $\mathcal{X}_{\star}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\min_{y\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ , $\mathcal{Y}_{\star}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\max_{x\in\mathchoice{\includegraphics[height=4.21 94pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[he ight=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{n}}\langle x,Ay\rangle$ , $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ , $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ , and $k:=\max\{k_{1},k_{2}\}$ . First, in Section 5.1, we design an algorithm that requires nearly $n\cdot\text{poly}(k)$ queries to identify $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ with high probability. We complement this result by showing in Section 5.2 that $\Omega(nk)$ queries are required in expectation to identify $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ even when $\min\{k_{1},k_{2}\}=1$ .
### 5.1 Randomized Algorithm
We begin with restating our main result. See 1
Here is the outline of our randomized algorithm. We begin by designing a randomized procedure called SetV in Section 5.1.1, which takes the values $k_{1}$ and $k_{2}$ as input and finds a set $\mathcal{V}$ of value of the game candidates that includes $V_{A}^{\star}$ . Next, we design a randomized procedure called Nash in Section 5.1.2 that also takes the values $k_{1}$ and $k_{2}$ as input, calls the procedure SetV, constructs a high dimensional matrix and finds its strict PSNE which maps to $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . Finally, as we show in Section 5.1.3, it is straightforward to relax the assumptions on knowing $k_{1}$ and $k_{2}$ by sequentially trying all possible combinations, starting from $k_{1}=1$ and $k_{2}=1$ , and stopping when we identify $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . Throughout this process, we use the procedure FindPsne to either find the strict PSNE or compute the set $\mathcal{V}$ using appropriate query oracles, ensuring that the query complexity remains nearly $n\cdot\text{poly}(k)$ .
#### 5.1.1 Finding the value of the game candidates with the knowledge of $k_{1}$ and $k_{2}$
This section involves three key steps: defining a large matrix, defining the query oracle, and identifying a set $\mathcal{V}$ of value of the game candidates.
We begin by assuming that we know the values $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ and $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ . Let $A\in\mathbb{R}^{n\times n}$ be the input matrix. We construct a matrix $A^{(k_{1},k_{2})}\in\mathbb{R}^{\binom{n}{k_{1}}\times\binom{n}{k_{2}}}$ as follows. First, we consider two bijections $f_{1}:[\binom{n}{k_{1}}]\rightarrow\binom{[n]}{k_{1}}$ and $f_{2}:[\binom{n}{k_{2}}]\rightarrow\binom{[n]}{k_{2}}$ . Next, we consider indices $i\in[\binom{n}{k_{1}}]$ and $j\in[\binom{n}{k_{2}}]$ . Let $M$ be the submatrix of $A$ with row indices $f_{1}(i)$ and column indices $f_{2}(j)$ . We now define $A_{i,j}^{(k_{1},k_{2})}:=V_{M}^{\star}$ . Now we prove the following lemma.
**Lemma 8**
*Let $i_{\star}:=f_{1}^{-1}(\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{ supp}(x))$ and $j_{\star}:=f_{2}^{-1}(\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{ supp}(y))$ . Then $(i_{\star},j_{\star})$ is a PSNE of $A^{(k_{1},k_{2})}$ and $A^{(k_{1},k_{2})}_{i_{\star},j_{\star}}=V_{A}^{\star}$ .*
* Proof*
For the sake of simplicity in presentation, let $\widehat{A}$ denote the matrix $A^{(k_{1},k_{2})}$ . Recall that $\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{\mathcal{I}}$ represents the simplex over a set of indices $\mathcal{I}$ . Let $(x_{\star},y_{\star})$ be a Nash equilibrium of $A$ such that $\operatorname*{supp}(x_{\star})=\bigcup\limits_{x\in\mathcal{X}_{\star}} \operatorname*{supp}(x)$ and $\operatorname*{supp}(y_{\star})=\bigcup\limits_{y\in\mathcal{Y}_{\star}} \operatorname*{supp}(y)$ . Such an equilibrium exists due to the fact that convex combination of multiple Nash equilibria is also a Nash equilibrium. Let $\widehat{x}\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0 pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{1}(i_{ \star})},\widehat{y}\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{f_{2}(j_{\star})}$ be strategies such that for all $i\in f_{1}(i_{\star})$ we have $\widehat{x}_{i}=(x_{\star})_{i}$ and for all $j\in f_{2}(j_{\star})$ we have $\widehat{y}_{j}=(y_{\star})_{j}$ . First we claim that $\widehat{A}_{i_{\star},j_{\star}}=V_{A}^{\star}$ . Consider the submatrix $M$ of $A$ with row indices $f_{1}(i_{\star})$ and column indices $f_{2}(j_{\star})$ . As $(x_{\star},y_{\star})$ is a Nash equilibrium of $A$ , and $\operatorname*{supp}(x_{\star})=f_{1}(i_{\star})$ and $\operatorname*{supp}(y_{\star})=f_{2}(j_{\star})$ , we have $\langle x,M\widehat{y}\rangle=V_{A}^{\star}$ for all $x\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0 pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0. 0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{1}(i_{\star})}$ and $\langle\widehat{x},My\rangle=V_{A}^{\star}$ for all $y\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0 pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0. 0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{2}(j_{\star})}$ . Hence $(\widehat{x},\widehat{y})$ is the Nash equilibrium of $M$ which implies that $\widehat{A}_{i_{\star},j_{\star}}=V_{M}^{\star}=\langle\widehat{x},M\widehat{y }\rangle=V_{A}^{\star}$ . Next consider an index $i\in[\binom{n}{k_{1}}]$ . Now we claim that $\widehat{A}_{i,j_{\star}}\leq V_{A}^{\star}$ . Consider the submatrix $M$ of $A$ with row indices $f_{1}(i)$ and column indices $f_{2}(j_{\star})$ . Now we have the following:
$$
\widehat{A}_{i,j_{\star}}=V_{M}^{\star}=\min_{y\in\mathchoice{\includegraphics
[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]
{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.
0pt 0.0pt]{graphics/simplex.pdf}}_{f_{2}(j_{\star})}}\max_{i\in f_{1}(i)}
\langle e_{i},My\rangle\leq\max_{i\in f_{1}(i)}\langle e_{i},M\widehat{y}
\rangle\leq V_{A}^{\star}.
$$
The last inequality follows from the definition of $\widehat{y}$ and the fact that $\max_{i\in[n]}\langle e_{i},Ay_{\star}\rangle=V_{A}^{\star}$ . Similarly we can show that for any $j\in[\binom{n}{k_{2}}]$ , we have $\widehat{A}_{i_{\star},j}\geq V_{A}^{\star}$ . Hence, $(i_{\star},j_{\star})$ is a PSNE. ∎
Now we describe the query oracle to observe the rows and columns of $A^{(k_{1},k_{2})}$ .
Query oracle $\mathcal{Q}$ to observe row $\hat{i}$ and column $\hat{j}$ of $A^{(k_{1},k_{2})}$ : Recall that $A^{(k_{1},k_{2})}_{i,j}$ only depends on the entries in the submatrix $M$ with row indices $f_{1}(i)$ and column indices $f_{2}(j)$ . Hence, if we want to observe the values in the set $\{A^{(k_{1},k_{2})}_{\hat{i},j}:j\in\{1,2,\ldots,\binom{n}{k_{2}}\}\}$ , then it suffices to query the entries of $A$ from the set $S_{1}=\{(i,j):i\in f_{1}(\hat{i}),j\in[n]\}$ . Since $|f_{1}(\hat{i})|=k_{1}$ , we have that $|S_{1}|=k_{1}\cdot n$ . Analogously, if we want to observe the values in the set $\{A^{(k_{1},k_{2})}_{i,\hat{j}}:i\in\{1,2,\ldots,\binom{n}{k_{1}}\}\}$ , then it suffices to query the entries of $A$ from the set $S_{2}=\{(i,j):i\in[n],j\in f_{2}(\hat{j})\}$ . Since $|f_{2}(\hat{j})|=k_{2}$ , we have $|S_{2}|=k_{2}\cdot n$ .
Now we describe SetV, a randomized algorithm that aims to find a set $\mathcal{V}$ that includes $V_{A}^{\star}$ .
Description of the randomized procedure SetV $(A,\delta,k_{1},k_{2})$ : Let $A\in\mathbb{R}^{n\times n}$ be the input matrix. Recall the construction of the matrix $A^{(k_{1},k_{2})}\in\mathbb{R}^{\binom{n}{k_{1}}\times\binom{n}{k_{2}}}$ from above. We now call the procedure FindPsne ( $A^{(k_{1},k_{2})},\delta$ ), where we use the query oracle $\mathcal{Q}$ to observe a row or column of $A^{(k_{1},k_{2})}$ . Let $\mathcal{V}$ be the set of values returned by FindPsne ( $A^{(k_{1},k_{2})},\delta$ ). We return the set $\mathcal{V}$ .
Now, we prove the following lemma regarding the procedure SetV.
**Lemma 9**
*Consider a matrix game $A\in\mathbb{R}^{n\times n}$ . Now we have the following:
- The procedure SetV $(A,\delta,k_{1},k_{2})$ queries at most $O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of $A$ .
- With probability at least $1-\delta$ , the procedure SetV $(A,\delta,k_{1},k_{2})$ returns a set $\mathcal{V}$ of distinct real numbers such that $V_{A}^{\star}\in\mathcal{V}$ .*
* Proof*
Due to Lemma 5, the number of calls made by FindPsne ( $A^{(k_{1},k_{2})},\delta$ ) to the query oracle $\mathcal{Q}$ is at most $O(\log(n^{k})\cdot\log(\frac{n^{k}}{\delta}))\leq O(k^{2}\cdot\log(n)\cdot\log (\frac{n}{\delta}))$ . Each call to the query oracle $\mathcal{Q}$ takes $O(nk)$ queries from the input matrix $A$ . Hence, the procedure FindPsne ( $A^{(k_{1},k_{2})},\delta$ ) queries at most $O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of $A$ and returns a set $\mathcal{V}$ . Due to Lemmas 7 and 8, with probability at least $1-\delta$ , $V_{A}^{\star}\in\mathcal{V}$ . ∎
#### 5.1.2 Algorithm for finding all Nash equilibria with the knowledge of $k_{1}$ and $k_{2}$
This section involves three key steps: modifying the large matrix from the previous section, defining the query oracle, and identifying the set of all Nash equilibria.
We begin by assuming that we know the values $k_{1}:=|\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{supp}(x)|$ and $k_{2}:=|\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{supp}(y)|$ . Let $A\in\mathbb{R}^{n\times n}$ be the input matrix. Let $\mathcal{V}$ be a set of distinct real numbers. Recall that $\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444pt,t rim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{\mathcal{I}}$ represents the simplex over a set of indices $\mathcal{I}$ . We construct a matrix $A^{(k_{1},k_{2},\mathcal{V})}\in\mathbb{R}^{\binom{n}{k_{1}}\times\binom{n}{k_ {2}}}$ as follows. First, we consider two bijections $f_{1}:[\binom{n}{k_{1}}]\rightarrow\binom{[n]}{k_{1}}$ and $f_{2}:[\binom{n}{k_{2}}]\rightarrow\binom{[n]}{k_{2}}$ . Next, we consider indices $i\in[\binom{n}{k_{1}}]$ and $j\in[\binom{n}{k_{2}}]$ . Let $M$ be the submatrix of $A$ with row indices $f_{1}(i)$ and column indices $f_{2}(j)$ . Let $\mathcal{X}_{M}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194pt,tr im=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4. 2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[ height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{f_{1}(i)}}\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{f_{2}(j)}}\langle x,My\rangle$ and $\mathcal{Y}_{M}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,tr im=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4. 2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[ height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{f_{2}(j)}}\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{f_{1}(j)}}\langle x,My\rangle$ . Let $k_{1,M}:=|\bigcup\limits_{x\in\mathcal{X}_{M}}\operatorname*{supp}(x)|$ and $k_{2,M}:=|\bigcup\limits_{y\in\mathcal{Y}_{M}}\operatorname*{supp}(y)|$ . If $V_{M}^{\star}\notin\mathcal{V}$ then we define $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}:=V_{M}^{\star}$ . If $V_{M}^{\star}\in\mathcal{V}$ then we define $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}$ as follows:
- If $k_{1,M}=k_{1}$ and $k_{2,M}=k_{2}$ then $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}=V_{M}^{\star}$
- If $k_{1,M}\neq k_{1}$ and $k_{2,M}\neq k_{2}$ then $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}=V_{M}^{\star}$
- If $k_{1,M}=k_{1}$ and $k_{2,M}\neq k_{2}$ then $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}=V_{M}^{\star}+\frac{\varepsilon}{4}$
- If $k_{1,M}\neq k_{1}$ and $k_{2,M}=k_{2}$ then $A_{i,j}^{(k_{1},k_{2},\mathcal{V})}=V_{M}^{\star}-\frac{\varepsilon}{4}$
where $\varepsilon=\min\limits_{v_{1},v_{2}\in\mathcal{V}:v_{1}\neq v_{2}}|v_{1}-v_{2}|$ . If $|\mathcal{V}|=1$ , then $\varepsilon=1$ . Now we prove the following lemma.
**Lemma 10**
*Let $i_{\star}:=f_{1}^{-1}(\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{ supp}(x))$ and $j_{\star}:=f_{2}^{-1}(\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{ supp}(y))$ . If $V_{A}^{\star}\in\mathcal{V}$ , then $(i_{\star},j_{\star})$ is the only strict PSNE of $A^{(k_{1},k_{2},\mathcal{V})}$ .*
* Proof*
Let us assume that $V_{A}^{\star}\in\mathcal{V}$ . For the sake of simplicity in presentation, let $\widehat{A}$ denote the matrix $A^{(k_{1},k_{2},\mathcal{V})}$ . Now due to Lemma 4 it suffices to show that $\widehat{A}_{i,j_{\star}}<\widehat{A}_{i_{\star},j_{\star}}<\widehat{A}_{i_{ \star},j}$ for all $i\in[\binom{n}{k_{1}}]\setminus\{i_{\star}\}$ and $j\in[\binom{n}{k_{2}}]\setminus\{j_{\star}\}$ . Let $(x_{\star},y_{\star})$ be a Nash equilibrium of $A$ such that the following holds:
- $\operatorname*{supp}(x_{\star})=\bigcup\limits_{x\in\mathcal{X}_{\star}} \operatorname*{supp}(x)$ and $\operatorname*{supp}(y_{\star})=\bigcup\limits_{y\in\mathcal{Y}_{\star}} \operatorname*{supp}(y)$
- For all $i\in\operatorname*{supp}(x_{\star})$ , we have $\langle e_{i},Ay_{\star}\rangle=V_{A}^{\star}$ , and for all $i\notin\operatorname*{supp}(x_{\star})$ , we have $\langle e_{i},Ay_{\star}\rangle<V_{A}^{\star}$
- For all $j\in\operatorname*{supp}(y_{\star})$ , we have $\langle x_{\star},Ae_{j}\rangle=V_{A}^{\star}$ , and for all $j\notin\operatorname*{supp}(y_{\star})$ , we have $\langle x_{\star},Ae_{j}\rangle>V_{A}^{\star}$
Such an equilibrium exists due to Lemma 1 and the fact that convex combination of multiple Nash equilibria is also a Nash equilibrium. Let $\widehat{x}\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0 pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{1}(i_{ \star})},\widehat{y}\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{f_{2}(j_{\star})}$ be strategies such that for all $i\in f_{1}(i_{\star})$ we have $\widehat{x}_{i}=(x_{\star})_{i}$ and for all $j\in f_{2}(j_{\star})$ we have $\widehat{y}_{j}=(y_{\star})_{j}$ . First we claim that $\widehat{A}_{i_{\star},j_{\star}}=V_{A}^{\star}$ . Consider the submatrix $M$ of $A$ with row indices $f_{1}(i_{\star})$ and column indices $f_{2}(j_{\star})$ . As $\operatorname*{supp}(x_{\star})=f_{1}(i_{\star})$ and $\operatorname*{supp}(y_{\star})=f_{2}(j_{\star})$ , we have $\langle x,M\widehat{y}\rangle=V_{A}^{\star}$ for all $x\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0 pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0. 0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{1}(i_{\star})}$ and $\langle\widehat{x},My\rangle=V_{A}^{\star}$ for all $y\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0pt 30.0pt 0.0pt 0.0 pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.73611pt,trim=0. 0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=3.44444 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{f_{2}(j_{\star})}$ . Hence $(\widehat{x},\widehat{y})$ is the Nash equilibrium of $M$ which implies that $V_{M}^{\star}=\langle\widehat{x},M\widehat{y}\rangle=V_{A}^{\star}$ . As $|\operatorname*{supp}(\widehat{x})|=k_{1}$ and $|\operatorname*{supp}(\widehat{y})|=k_{2}$ we have $\widehat{A}_{i_{\star},j_{\star}}=V_{A}^{\star}$ . Next consider an index $i\in[\binom{n}{k_{1}}]$ such that $f_{1}(i)\cap f_{1}(i_{\star})=\emptyset$ . Now we claim that $\widehat{A}_{i,j_{\star}}<V_{A}^{\star}$ . Consider the submatrix $M$ of $A$ with row indices $f_{1}(i)$ and column indices $f_{2}(j_{\star})$ . Now we have the following:
$$
V_{M}^{\star}=\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0
pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt
,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height
=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}_{f_{2}(j_{\star})}}\max_{i\in f_{1}(i)}\langle e_{i},My\rangle
\leq\max_{i\in f_{1}(i)}\langle e_{i},M\widehat{y}\rangle<V_{A}^{\star}.
$$
The last inequality follows from our choice of $(x_{\star},y_{\star})$ above and $f_{1}(i)\cap f_{1}(i_{\star})=\emptyset$ . If $V_{M}^{\star}\notin\mathcal{V}$ , then $\widehat{A}_{i,j_{\star}}=V_{M}^{\star}<V_{A}^{\star}$ . If $V_{M}^{\star}\in\mathcal{V}$ , then $\widehat{A}_{i,j_{\star}}\leq V_{M}^{\star}+\frac{\varepsilon}{4}<V_{A}^{\star}$ . The last inequality follows from the definition $\varepsilon$ . Analogously, we can show that for any $j\in[\binom{n}{k_{2}}]$ such that $f_{2}(j)\cap f_{2}(j_{\star})=\emptyset$ , we have $\widehat{A}_{i_{\star},j}>V_{A}^{\star}$ . Next consider an index $i\in[\binom{n}{k_{1}}]\setminus\{i_{\star}\}$ such that $f_{1}(i)\cap f_{1}(i_{\star})\neq\emptyset$ . Now we claim that $\widehat{A}_{i,j_{\star}}<V_{A}^{\star}$ . Consider the submatrix $M$ of $A$ with row indices $f_{1}(i)$ and column indices $f_{2}(j_{\star})$ . Now we have the following:
$$
V_{M}^{\star}=\min_{y\in\mathchoice{\includegraphics[height=4.2194pt,trim=0.0
pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=4.2194pt
,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height
=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{
\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/
simplex.pdf}}_{f_{2}(j_{\star})}}\max_{i\in f_{1}(i)}\langle e_{i},My\rangle
\leq\max_{i\in f_{1}(i)}\langle e_{i},M\widehat{y}\rangle=V_{A}^{\star}.
$$
Recall the definition of $\mathcal{X}_{M},\mathcal{Y}_{M},k_{1,M}$ and $k_{2,M}$ . If $V_{M}^{\star}<V_{A}^{\star}$ and $V_{M}^{\star}\notin\mathcal{V}$ , then we have $\widehat{A}_{i,j_{\star}}=V_{M}^{\star}<V_{A}^{\star}$ . If $V_{M}^{\star}<V_{A}^{\star}$ and $V_{M}^{\star}\in\mathcal{V}$ , then we have $\widehat{A}_{i,j_{\star}}\leq V_{M}^{\star}+\frac{\varepsilon}{4}<V_{A}^{\star}$ . Let us consider the case when $V_{M}^{\star}=V_{A}^{\star}$ . In this case, we have $\widehat{y}\in\mathcal{Y}_{M}$ which implies that $k_{2,M}=k_{2}$ . Consider $x\in\mathcal{X}_{M}$ . For any $i\notin f_{1}(i)\cap f_{1}(i_{\star})$ we have $\langle e_{i},M\widehat{y}\rangle<V_{A}^{\star}$ . This implies that $\operatorname*{supp}(x)\subseteq f_{1}(i)\cap f_{1}(i_{\star})$ . As $x$ was chosen arbitrarily, we have $\bigcup\limits_{x\in\mathcal{X}_{M}}\operatorname*{supp}(x)\subseteq f_{1}(i) \cap f_{1}(i_{\star})$ . As $f_{1}(i)\neq f_{1}(i_{\star})$ this implies $|f_{1}(i)\cap f_{1}(i_{\star})|<k_{1}$ . This implies that $k_{1,M}\neq k_{1}$ . Hence due to our construction, we have $\widehat{A}_{i,j_{\star}}=V_{A}^{\star}-\frac{\varepsilon}{4}<V_{A}^{\star}$ . Analogously, we can show that for any $j\in[\binom{n}{k_{2}}]\setminus\{j_{\star}\}$ such that $f_{2}(j)\cap f_{2}(j_{\star})\neq\emptyset$ , we have $\widehat{A}_{i_{\star},j}>V_{A}^{\star}$ . Combining all the cases above, we have $\widehat{A}_{i,j_{\star}}<\widehat{A}_{i_{\star},j_{\star}}<\widehat{A}_{i_{ \star},j}$ for all $i\in[\binom{n}{k_{1}}]\setminus\{i_{\star}\}$ and $j\in[\binom{n}{k_{2}}]\setminus\{j_{\star}\}$ . ∎
Now we describe the query oracle to observe the rows and columns of $A^{(k_{1},k_{2},\mathcal{V})}$ .
Query oracle $\mathcal{Q}$ to observe row $\hat{i}$ and column $\hat{j}$ of $A^{(k_{1},k_{2},\mathcal{V})}$ : Recall that $A^{(k_{1},k_{2},\mathcal{V})}_{i,j}$ only depends on the set $\mathcal{V}$ and the entries in the submatrix $M$ with row indices $f_{1}(i)$ and column indices $f_{2}(j)$ . Hence, if we want to observe the values in the set $\{A^{(k_{1},k_{2},\mathcal{V})}_{\hat{i},j}:j\in\{1,2,\ldots,\binom{n}{k_{2}}\}\}$ , then it suffices to query the entries of $A$ from the set $S_{1}=\{(i,j):i\in f_{1}(\hat{i}),j\in[n]\}$ . Since $|f_{1}(\hat{i})|=k_{1}$ , we have that $|S_{1}|=k_{1}\cdot n$ . Analogously, if we want to observe the values in the set $\{A^{(k_{1},k_{2},\mathcal{V})}_{i,\hat{j}}:i\in\{1,2,\ldots,\binom{n}{k_{1}}\}\}$ , then it suffices to query the entries of $A$ from the set $S_{2}=\{(i,j):i\in[n],j\in f_{2}(\hat{j})\}$ . Since $|f_{2}(\hat{j})|=k_{2}$ , we have $|S_{2}|=k_{2}\cdot n$ .
Now we describe Nash, a randomized algorithm that finds $\mathcal{X}_{\star}$ and $\mathcal{Y}_{\star}$ when provided with the values $k_{1}$ and $k_{2}$ .
Description of the randomized procedure Nash $(A,\delta,k_{1},k_{2})$ : Let $A\in\mathbb{R}^{n\times n}$ be the input matrix. First, we call the procedure SetV $(A,\delta/2,k_{1},k_{2})$ . Let $\mathcal{V}$ be the set returned by this procedure. Recall the construction of the matrix $A^{(k_{1},k_{2},\mathcal{V})}\in\mathbb{R}^{\binom{n}{k_{1}}\times\binom{n}{k_ {2}}}$ from above. We now call the procedure FindPsne ( $A^{(k_{1},k_{2},\mathcal{V})},\delta/2$ ), where we use the query oracle $\mathcal{Q}$ to observe a row or column of $A^{(k_{1},k_{2},\mathcal{V})}$ .
Let $(\hat{i},\hat{j})$ be the pair returned by the procedure FindPsne $(A^{(k_{1},k_{2},\mathcal{V})},\delta/2)$ . Let $M_{1}$ be the submatrix of $A$ with row indices $f_{1}(\hat{i})$ and column indices $[n]$ , and $M_{2}$ be the submatrix of $A$ with row indices $[n]$ and column indices $f_{2}(\hat{j})$ . Let $\mathcal{X}_{M_{1}}:=\arg\max_{x\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{f_{1}(\hat{i})}}\min_{y\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,M_{1}y\rangle$ and $\mathcal{Y}_{M_{2}}:=\arg\min_{y\in\mathchoice{\includegraphics[height=4.2194 pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt] {graphics/simplex.pdf}}_{f_{2}(\hat{j})}}\max_{x\in\mathchoice{ \includegraphics[height=4.2194pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}{\includegraphics[height=4.2194pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{ graphics/simplex.pdf}}{\includegraphics[height=3.31529pt,trim=0.0pt 30.0pt 0.0 pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=2.41112pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}_{n}}\langle x,M_{2}y\rangle$ . For any vector $v$ and the set of indices $\mathcal{S}=\{i_{1},\ldots,i_{\ell}\}$ , let $v_{\mathcal{S}}$ denote the vector $(v_{i_{1}},\ldots,v_{i_{\ell}})$ . Let $\mathcal{X}_{A}:=\{x\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{n}:x_{f_{1}(\hat{i})}\in\mathcal{X}_{M_{1}}\}$ and $\mathcal{Y}_{A}:=\{y\in\mathchoice{\includegraphics[height=6.02773pt,trim=0.0 pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[height=6.02773 pt,trim=0.0pt 25.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{\includegraphics[heig ht=4.73611pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/simplex.pdf}}{ \includegraphics[height=3.44444pt,trim=0.0pt 30.0pt 0.0pt 0.0pt]{graphics/ simplex.pdf}}_{m}:y_{f_{2}(\hat{j})}\in\mathcal{Y}_{M_{2}}\}$ . We return $\mathcal{X}_{A}\times\mathcal{Y}_{A}$ as the set of all Nash equilibria if the following conditions hold:
1. $V_{M_{1}}^{\star}=V_{M_{2}}^{\star}$ .
1. $\bigcap\limits_{x\in\mathcal{X}_{A}}I_{x}^{\prime}=f_{2}(\hat{j})$ where $I_{x}^{\prime}=\{j\in[n]:V_{M_{1}}^{\star}=\langle x,Ae_{j}\rangle\}$ .
1. $\bigcap\limits_{y\in\mathcal{Y}_{A}}I_{y}^{\prime}=f_{1}(\hat{i})$ where $I_{y}^{\prime}=\{i\in[n]:V_{M_{2}}^{\star}=\langle e_{i},Ay\rangle\}$ .
The above conditions can be checked by querying all the entries of $A$ in the set $\{(i,j):i\in f_{1}(\hat{i}),j\in[n]\}\cup\{(i,j):i\in[n],j\in f_{2}(\hat{j})\}$ . If any one of the conditions above does not hold, we return "failure" instead.
Now we state the guarantees of our randomized procedure.
**Lemma 11**
*Consider a matrix game $A\in\mathbb{R}^{n\times n}$ . Now we have the following:
- The procedure Nash $(A,\delta,k_{1},k_{2})$ queries at most $O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of $A$ .
- With probability at least $1-\delta$ , the procedure Nash $(A,\delta,k_{1},k_{2})$ returns $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ .*
* Proof*
Due to Lemma 9, SetV $(A,\delta/2,k_{1},k_{2})$ queries at most $O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of $A$ . Next due to Lemma 5, the number of calls made by FindPsne ( $A^{(k_{1},k_{2},\mathcal{V})},\delta/2$ ) to the query oracle $\mathcal{Q}$ is at most $O(\log(n^{k})\cdot\log(\frac{n^{k}}{\delta}))\leq O(k^{2}\cdot\log(n)\cdot\log (\frac{n}{\delta}))$ . Each call to the query oracle $\mathcal{Q}$ takes $O(nk)$ queries from the input matrix $A$ . Moreover, to check the conditions (C1), (C2) and (C3) it takes $O(nk)$ queries. Hence, the procedure Nash $(A,\delta,k_{1},k_{2})$ queries at most $O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))$ entries of $A$ . Due Lemma 9, with probability at least $1-\frac{\delta}{2}$ , SetV $(A,\delta/2,k_{1},k_{2})$ returns a set $\mathcal{V}$ such that $V_{A}^{\star}\in\mathcal{V}$ . Recall the construction of the sets $\mathcal{X}_{A}$ and $\mathcal{Y}_{A}$ . Due to Lemma 10 and Lemma 6, the procedure FindPsne $(A^{(k_{1},k_{2},\mathcal{V})},\delta/2)$ returns the pair $(i_{\star},j_{\star})$ with probability at least $1-\frac{\delta}{2}$ if $V_{A}^{\star}\in\mathcal{V}$ , where $i_{\star}:=f_{1}^{-1}(\bigcup\limits_{x\in\mathcal{X}_{\star}}\operatorname*{ supp}(x))$ and $j_{\star}:=f_{2}^{-1}(\bigcup\limits_{y\in\mathcal{Y}_{\star}}\operatorname*{ supp}(y))$ . The previous facts along with Lemma 3 imply that with probability at least $1-\delta$ , we have $\mathcal{X}_{A}=\mathcal{X}_{\star}$ and $\mathcal{Y}_{A}=\mathcal{Y}_{\star}$ . As the set $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ satisfies the conditions (C1), (C2) and (C3) due to Lemma 3, with probability at least $1-\delta$ , the procedure Nash $(A,\delta,k_{1},k_{2})$ returns $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . ∎
#### 5.1.3 Relaxing the assumption on the knowledge of $k_{1}$ and $k_{2}$ .
It is straightforward to relax the assumption on the knowledge of $k_{1}$ and $k_{2}$ . Beginning with $s=1$ , we call the procedure Nash ( $A,\delta,s_{1},s_{2}$ ) on all possible pairs $(s_{1},s_{2})$ such that $\max\{s_{1},s_{2}\}=s$ . Note that the number of such pairs is $2s-1$ . If for any one of the pairs $(s_{1},s_{2})$ the procedure Nash ( $A,\delta,s_{1},s_{2}$ ) returns a set $\mathcal{X}_{A}\times\mathcal{Y}_{A}$ , we return it as the set of all Nash equilibria and terminate the process. If for all the pairs $(s_{1},s_{2})$ the procedure Nash ( $A,\delta,s_{1},s_{2}$ ) returns "failure" then we increment $s$ by one and repeat the process. With probability $1-\delta$ , this process terminates at $s=k$ and returns the set $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ . This is because the set $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ is the only set that satisfies the conditions (C1), (C2) and (C3) due to Lemma 3. Hence, with probability $1-\delta$ , the total number of queries taken from $A$ is at most $(\sum_{s=1}^{k}2s-1)\cdot O(nk^{3}\cdot\log n\cdot\log(\frac{n}{\delta}))=O(nk ^{5}\cdot\log n\cdot\log(\frac{n}{\delta}))$ .
### 5.2 Lower bound
In this section, we prove the following lower bound result on the query complexity.
See 2
* Proof*
For each $\ell\in[n]$ , let $\mathcal{I}_{\ell}$ be the set of all $n\times n$ matrices such that the following holds:
1. all entries $(i,j)$ have a value two for $i\in[n]$ and $j\in[n]\setminus[k]$
1. for each $i\in[n]$ such that $i\notin\{1\}\cup\{\ell\}$ , there exists exactly one index $j\in[k]$ such that the entry $(i,j)$ has a value of zero.
1. rest of the entries of the matrix have a value of one.
Fix an index $\ell\in[n]$ . For any matrix in $\mathcal{I}_{\ell}$ , it can be easily shown using Lemma 1 that $\mathcal{S}_{\ell}:=\{(e_{i},e_{j}):i\in\{1\}\cup\{\ell\},j\in[k]\}$ is the set of all pure strategy nash equilibria, and that their convex combination is the set of all Nash equilibria. Also observe that for two distinct indices $\ell_{1},\ell_{2}\in[n]$ , we have $\mathcal{S}_{\ell_{1}}\neq\mathcal{S}_{\ell_{2}}$ . Let $\mathcal{D}$ be a uniform distribution over the set of matrices in $\mathcal{I}_{1}$ . Observe that for any matrix $A\in\mathcal{I}_{1}$ , we have $\max\{k_{1},k_{2}\}=k$ and $\min\{k_{1},k_{2}\}=1$ . Let us assume that the randomized algorithm $\mathcal{A}$ is a distribution over a set $\mathcal{T}$ of deterministic algorithms. Fix a deterministic instance $\operatorname{\mathsf{Alg}}\in\mathcal{T}$ . Let $\tau(\operatorname{\mathsf{Alg}},A)$ denote the number of queries taken by $\operatorname{\mathsf{Alg}}$ from an input matrix $A$ . We now aim to show that $\mathbb{E}_{A\sim\mathcal{D}}[\tau(\operatorname{\mathsf{Alg}},A)]\geq\frac{(n -1)(k+1)}{2}$ . For the sake of simplicity in presentation, let us assume that $\operatorname{\mathsf{Alg}}$ does not query an entry more than once. Consider an input matrix $A\in\mathcal{I}_{1}$ . As $\operatorname{\mathsf{Alg}}$ correctly computes the set of all PSNEs of the input matrix, $\operatorname{\mathsf{Alg}}$ has to query all those entries of $A$ that have a value of $0 0$ . For the sake of contradiction, let us assume that $\operatorname{\mathsf{Alg}}$ does not query an entry $(i^{\prime},j^{\prime})$ that has a value of $0 0$ . One can create a matrix $B\in\mathcal{I}_{i^{\prime}}$ such that for all $(i,j)\neq(i^{\prime},j^{\prime})$ we have $B_{i,j}=A_{i,j}$ and $B_{i^{\prime},j^{\prime}}=1$ . As $\operatorname{\mathsf{Alg}}$ terminates without querying the entry $(i^{\prime},j^{\prime})$ it cannot distinguish between $A$ and $B$ and hence outputs the wrong set of PSNEs for at least one of the matrices which is contradiction to our assumption. Hence, all the $n-1$ zero-valued entries are always queried for all matrices in $\mathcal{I}_{1}$ . We continue to focus on the input matrices $A$ in $\mathcal{I}_{1}$ . Fix $i\in[n]\setminus\{1\}$ . Let $X_{i,A}$ denote the number of entries of the set $\{(i,j):j\in[k]\}$ that $\operatorname{\mathsf{Alg}}$ queries before querying the zero-valued entry in row $i$ of an input matrix $A$ . Let $\mathcal{F}_{i}$ be the family of functions $f:[n]\setminus\{1,i\}\rightarrow[k]$ . Fix $f\in\mathcal{F}_{i}$ . Let $\mathcal{I}_{1,f}$ be the largest subset of $\mathcal{I}_{1}$ such that for all $A\in\mathcal{I}_{1,f}$ we have $A_{\hat{i},f(\hat{i})}=0$ for all $\hat{i}\in[n]\setminus\{1,i\}$ . Let $(i,j_{1})$ be the first entry to be queried from the set $\{(i,j):j\in[k]\}$ by $\operatorname{\mathsf{Alg}}$ if input matrix $A$ is chosen from the set $\mathcal{I}_{1,f}$ . Recursively for $\ell\geq 2$ , let $(i,j_{\ell})$ be the $\ell$ -th entry to be queried from the set $\{(i,j):j\in[k]\}$ by $\operatorname{\mathsf{Alg}}$ if input matrix $A$ is chosen from the set $\mathcal{I}_{1,f}$ and the first $\ell-1$ entries queried from the set $\{(i,j):j\in[k]\}$ have a value of one. Note that the sequence $j_{1},\ldots,j_{k}$ is fixed as $\operatorname{\mathsf{Alg}}$ behaves identically on all the matrices in $\mathcal{I}_{1,f}$ until it queries the entry in row $i$ that has a value of $0 0$ . Fix $m\in\{0,\ldots,k-1\}$ . Now observe that $\mathbb{P}_{A\sim\mathcal{D}}[X_{i,A}=m|A\in\mathcal{I}_{1,f}]=\mathbb{P}_{A \sim\mathcal{D}}[A_{i,j_{m+1}}=0|A\in\mathcal{I}_{1,f}]=\frac{1}{k}$ as $|\mathcal{I}_{1,f}|=k$ and there is exactly one matrix $A\in\mathcal{I}_{1,f}$ such that $A_{i,j_{m+1}}=0$ . Now we have the following:
$$
\displaystyle\mathbb{P}_{A\sim\mathcal{D}}[X_{i,A}=m] \displaystyle=\sum_{f\in\mathcal{F}_{i}}\mathbb{P}_{A\sim\mathcal{D}}[X_{i,A}=
m|A\in\mathcal{I}_{1,f}]\cdot\mathbb{P}_{A\sim\mathcal{D}}[A\in\mathcal{I}_{1,
f}] \displaystyle=\frac{1}{k}\cdot\sum_{f\in\mathcal{F}_{i}}\mathbb{P}_{A\sim
\mathcal{D}}[A\in\mathcal{I}_{1,f}] \displaystyle=\frac{1}{k}
$$ Hence, $\mathbb{E}_{A\sim\mathcal{D}}[X_{i,A}]=\sum_{m=0}^{k-1}\frac{m}{k}=\frac{k-1}{2}$ . Now observe that $\mathbb{E}_{A\sim\mathcal{D}}[\tau(\operatorname{\mathsf{Alg}},A)]\geq(n-1)+ \sum_{i=2}^{n}\mathbb{E}_{A\sim\mathcal{D}}[X_{i,A}]=\frac{(n-1)(k+1)}{2}$ . Now due to Yao’s lemma, we have that
$$
\max_{A\in\mathcal{I}_{1}}\operatorname*{\mathbb{E}}[\tau(A)]\geq\min_{
\operatorname{\mathsf{Alg}}\in\mathcal{T}}\operatorname*{\mathbb{E}}_{A\sim
\mathcal{D}}[\tau(\operatorname{\mathsf{Alg}},A)]\geq\frac{(n-1)(k+1)}{2}.
$$
∎
## 6 Conclusion and Open Questions
This paper characterizes the query complexity of finding the set of all Nash equilibria, $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ , in two-player zero-sum matrix games. In Theorem 1, we demonstrated that with probability at least $1-\delta$ , the set $\mathcal{X}_{\star}\times\mathcal{Y}_{\star}$ can be found by querying at most $O(nk^{5}\cdot\operatorname*{polylog}(\frac{n}{\delta}))$ entries. Additionally, in Theorem 2, we established a lower bound of $\Omega(nk)$ . Bridging the gap between this upper and lower bound remains an important open question.
Furthermore, note that our algorithm does not run in polynomial time. Therefore, designing a polynomial-time randomized algorithm that requires $o(n^{2})$ queries and finds a Nash equilibrium presents another compelling open problem. We conjecture that such an algorithm would need to be strongly polynomial.
Finally, it is worth recalling that there exists a deterministic algorithm with a query complexity of $O(n)$ to find the strict PSNE. The question of whether a deterministic algorithm can be developed to find the set of all Nash equilibria with $o(n^{2})$ queries is yet another important open question.
## ACKNOWLEDGEMENTS
This work was supported in part by NSF TRIPODS CCF Award #2023166 and a Northrop Grumman University Research Award. We thank Prof. Rahul Savani for his feedback on the paper.
## References
- Adler [2013] Ilan Adler. The equivalence of linear programs and zero-sum games. International Journal of Game Theory, 42(1):165, 2013.
- Auger et al. [2014] David Auger, Jialin Liu, Sylvie Ruette, David Lupien St-Pierre, and Olivier Teytaud. Sparse binary zero-sum games. In ACML, 2014.
- Babichenko [2016] Yakov Babichenko. Query complexity of approximate nash equilibria. Journal of the ACM (JACM), 63(4):1–24, 2016.
- Babichenko [2020] Yakov Babichenko. Informational bounds on equilibria (a survey). ACM SIGecom Exchanges, 17(2):25–45, 2020.
- Bienstock et al. [1991] Daniel Bienstock, Fan Chung, Michael L Fredman, Alejandro A Schäffer, Peter W Shor, and Subhash Suri. A note on finding a strict saddlepoint. The American mathematical monthly, 98(5):418–419, 1991.
- Bohnenblust et al. [1948] HF Bohnenblust, MA Girshick, RN Snow, Melvin Dresher, Olaf Helmer-Hirschberg, JCC McKinsey, Lloyd S Shapley, and Theodore Edward Harris. Mathematical theory of zero-sum two-person games with a finite number or a continuum of strategies. 1948.
- Bohnenblust et al. [1950] HF Bohnenblust, S Karlin, and LS Shapley. Solutions of discrete, two-person games. Contributions to the Theory of Games, 1:51–72, 1950.
- Brooks and Reny [2021] Benjamin Brooks and Philip J Reny. A canonical game–nearly 75 years in the making–showing the equivalence of matrix games and linear programming. Available at SSRN 3851583, 2021.
- Chen et al. [2009] Xi Chen, Xiaotie Deng, and Shang-Hua Teng. Settling the complexity of computing two-player nash equilibria. Journal of the ACM (JACM), 56(3):1–57, 2009.
- Dallant et al. [2024a] Justin Dallant, Frederik Haagensen, Riko Jacob, László Kozma, and Sebastian Wild. Finding the saddlepoint faster than sorting. In 2024 Symposium on Simplicity in Algorithms (SOSA), pages 168–178. SIAM, 2024a.
- Dallant et al. [2024b] Justin Dallant, Frederik Haagensen, Riko Jacob, László Kozma, and Sebastian Wild. An optimal randomized algorithm for finding the saddlepoint. arXiv preprint arXiv:2401.06512, 2024b.
- Dantzig [1951] George B Dantzig. A proof of the equivalence of the programming problem and the game problem. Activity analysis of production and allocation, 13, 1951.
- Daskalakis et al. [2007] Constantinos Daskalakis, Aranyak Mehta, and Christos Papadimitriou. Progress in approximate nash equilibria. In Proceedings of the 8th ACM Conference on Electronic Commerce, pages 355–358, 2007.
- Daskalakis et al. [2009] Constantinos Daskalakis, Aranyak Mehta, and Christos Papadimitriou. A note on approximate nash equilibria. Theoretical Computer Science, 410(17):1581–1588, 2009.
- Deligkas et al. [2022] Argyrios Deligkas, Michail Fasoulakis, and Evangelos Markakis. A polynomial-time algorithm for 1/3-approximate nash equilibria in bimatrix games. arXiv preprint arXiv:2204.11525, 2022.
- Deligkas et al. [2023] Argyrios Deligkas, Michail Fasoulakis, and Evangelos Markakis. A polynomial-time algorithm for 1/2-well-supported nash equilibria in bimatrix games. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 3777–3787. SIAM, 2023.
- Fearnley and Savani [2016] John Fearnley and Rahul Savani. Finding approximate nash equilibria of bimatrix games via payoff queries. ACM Transactions on Economics and Computation (TEAC), 4(4):1–19, 2016.
- Fearnley et al. [2013] John Fearnley, Martin Gairing, Paul Goldberg, and Rahul Savani. Learning equilibria of games via payoff queries. In Proceedings of the fourteenth ACM conference on Electronic commerce, pages 397–414, 2013.
- Goldberg and Roth [2016] Paul W Goldberg and Aaron Roth. Bounds for the query complexity of approximate equilibria. ACM Transactions on Economics and Computation (TEAC), 4(4):1–25, 2016.
- Karlin and Peres [2017] Anna R Karlin and Yuval Peres. Game theory, alive, volume 101. American Mathematical Soc., 2017.
- Maiti et al. [2024] Arnab Maiti, Ross Boczar, Kevin Jamieson, and Lillian Ratliff. Near-optimal pure exploration in matrix games: A generalization of stochastic bandits & dueling bandits. In International Conference on Artificial Intelligence and Statistics, pages 2602–2610. PMLR, 2024.
- v. Neumann [1928] J v. Neumann. Zur theorie der gesellschaftsspiele. Mathematische annalen, 100(1):295–320, 1928.