200

Circuit Complexity of Visual Search

International Symposium on Mathematical Foundations of Computer Science (MFCS), 2021
Abstract

We study computational hardness of feature and conjunction search through the lens of circuit complexity. Let x=(x1,...,xn)x = (x_1, ... , x_n) (resp., y=(y1,...,yn)y = (y_1, ... , y_n)) be Boolean variables each of which takes the value one if and only if a neuron at place ii detects a feature (resp., another feature). We then simply formulate the feature and conjunction search as Boolean functions FTRn(x)=i=1nxi{\rm FTR}_n(x) = \bigvee_{i=1}^n x_i and CONJn(x,y)=i=1nxiyi{\rm CONJ}_n(x, y) = \bigvee_{i=1}^n x_i \wedge y_i, respectively. We employ a threshold circuit or a discretized circuit (such as a sigmoid circuit or a ReLU circuit with discretization) as our models of neural networks, and consider the following four computational resources: [i] the number of neurons (size), [ii] the number of levels (depth), [iii] the number of active neurons outputting non-zero values (energy), and [iv] synaptic weight resolution (weight). We first prove that any threshold circuit CC of size ss, depth dd, energy ee and weight ww satisfies logrk(MC)ed(logs+logw+logn)\log rk(M_C) \le ed (\log s + \log w + \log n), where rk(MC)rk(M_C) is the rank of the communication matrix MCM_C of a 2n2n-variable Boolean function that CC computes. Since CONJn{\rm CONJ}_n has rank 2n2^n, we have ned(logs+logw+logn)n \le ed (\log s + \log w + \log n). Thus, an exponential lower bound on the size of even sublinear-depth threshold circuits exists if the energy and weight are sufficiently small. Since FTRn{\rm FTR}_n is computable independently of nn, our result suggests that computational capacity for the feature and conjunction search are different. We also show that the inequality is tight up to a constant factor if ed=o(n/logn)ed = o(n/ \log n). We next show that a similar inequality holds for any discretized circuit. Thus, if we regard the number of gates outputting non-zero values as a measure for sparse activity, our results suggest that larger depth helps neural networks to acquire sparse activity.

View on arXiv
Comments on this paper